mirror of
https://github.com/JakubMelka/PDF4QT.git
synced 2024-12-28 17:20:02 +01:00
356 lines
18 KiB
C++
356 lines
18 KiB
C++
// Copyright (C) 2018 Jakub Melka
|
|
//
|
|
// This file is part of PdfForQt.
|
|
//
|
|
// PdfForQt is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// PdfForQt is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with PDFForQt. If not, see <https://www.gnu.org/licenses/>.
|
|
|
|
|
|
#include <QtTest>
|
|
#include <QMetaType>
|
|
|
|
#include "pdfparser.h"
|
|
#include "pdfconstants.h"
|
|
#include "pdfflatmap.h"
|
|
|
|
#include <regex>
|
|
|
|
class LexicalAnalyzerTest : public QObject
|
|
{
|
|
Q_OBJECT
|
|
|
|
public:
|
|
explicit LexicalAnalyzerTest();
|
|
virtual ~LexicalAnalyzerTest() override;
|
|
|
|
private slots:
|
|
void test_null();
|
|
void test_numbers();
|
|
void test_strings();
|
|
void test_name();
|
|
void test_bool();
|
|
void test_ad();
|
|
void test_command();
|
|
void test_invalid_input();
|
|
void test_header_regexp();
|
|
void test_flat_map();
|
|
|
|
private:
|
|
void scanWholeStream(const char* stream);
|
|
void testTokens(const char* stream, const std::vector<pdf::PDFLexicalAnalyzer::Token>& tokens);
|
|
|
|
QString getStringFromTokens(const std::vector<pdf::PDFLexicalAnalyzer::Token>& tokens);
|
|
};
|
|
|
|
LexicalAnalyzerTest::LexicalAnalyzerTest()
|
|
{
|
|
|
|
}
|
|
|
|
LexicalAnalyzerTest::~LexicalAnalyzerTest()
|
|
{
|
|
|
|
}
|
|
|
|
void LexicalAnalyzerTest::test_null()
|
|
{
|
|
using Token = pdf::PDFLexicalAnalyzer::Token;
|
|
using Type = pdf::PDFLexicalAnalyzer::TokenType;
|
|
|
|
testTokens("null", { Token(Type::Null) });
|
|
testTokens(" null ", { Token(Type::Null), Token(Type::EndOfFile) });
|
|
testTokens("%null\n null %comment", { Token(Type::Null), Token(Type::EndOfFile) });
|
|
testTokens(" \n\t null\n", { Token(Type::Null), Token(Type::EndOfFile) });
|
|
testTokens(" null %and null\n null", { Token(Type::Null), Token(Type::Null) });
|
|
testTokens(" null %and null\n null ", { Token(Type::Null), Token(Type::Null), Token(Type::EndOfFile) });
|
|
}
|
|
|
|
void LexicalAnalyzerTest::test_numbers()
|
|
{
|
|
using Token = pdf::PDFLexicalAnalyzer::Token;
|
|
using Type = pdf::PDFLexicalAnalyzer::TokenType;
|
|
|
|
testTokens("1 +2 -3 +40 -55", { Token(Type::Integer, 1), Token(Type::Integer, 2), Token(Type::Integer, -3), Token(Type::Integer, 40), Token(Type::Integer, -55) });
|
|
testTokens(".0 0.1 3.5 -4. +5.0 -6.58 7.478", { Token(Type::Real, 0.0), Token(Type::Real, 0.1), Token(Type::Real, 3.5), Token(Type::Real, -4.0), Token(Type::Real, 5.0), Token(Type::Real, -6.58), Token(Type::Real, 7.478) });
|
|
testTokens("1000000000000000000000000000", { Token(Type::Real, 1e27) });
|
|
}
|
|
|
|
void LexicalAnalyzerTest::test_strings()
|
|
{
|
|
using Token = pdf::PDFLexicalAnalyzer::Token;
|
|
using Type = pdf::PDFLexicalAnalyzer::TokenType;
|
|
|
|
testTokens("(Simple string)", { Token(Type::String, QByteArray("Simple string")) });
|
|
testTokens("(String with (brackets))", { Token(Type::String, QByteArray("String with (brackets)")) });
|
|
testTokens("(String with \\( unbalanced brackets \\(\\))", { Token(Type::String, QByteArray("String with ( unbalanced brackets ()")) });
|
|
testTokens("()", { Token(Type::String, QByteArray("")) });
|
|
testTokens("(Text with special character: \\n)", { Token(Type::String, QByteArray("Text with special character: \n")) });
|
|
testTokens("(Text with special character: \\r)", { Token(Type::String, QByteArray("Text with special character: \r")) });
|
|
testTokens("(Text with special character: \\t)", { Token(Type::String, QByteArray("Text with special character: \t")) });
|
|
testTokens("(Text with special character: \\b)", { Token(Type::String, QByteArray("Text with special character: \b")) });
|
|
testTokens("(Text with special character: \\f)", { Token(Type::String, QByteArray("Text with special character: \f")) });
|
|
testTokens("(Text with special character: \\()", { Token(Type::String, QByteArray("Text with special character: (")) });
|
|
testTokens("(Text with special character: \\))", { Token(Type::String, QByteArray("Text with special character: )")) });
|
|
testTokens("(Text with special character: \\\\)", { Token(Type::String, QByteArray("Text with special character: \\")) });
|
|
testTokens("(\53)", { Token(Type::String, QByteArray("+")) });
|
|
testTokens("(\0533)", { Token(Type::String, QByteArray("+3")) });
|
|
testTokens("(\053)", { Token(Type::String, QByteArray("+")) });
|
|
testTokens("(\053053)", { Token(Type::String, QByteArray("+053")) });
|
|
testTokens("(\5)", { Token(Type::String, QByteArray("\5")) });
|
|
testTokens("<901FA3>", { Token(Type::String, QByteArray("\220\037\243")) });
|
|
testTokens("<901fa3>", { Token(Type::String, QByteArray("\220\037\243")) });
|
|
testTokens("<901fa>", { Token(Type::String, QByteArray("\220\037\240")) });
|
|
testTokens("<901FA>", { Token(Type::String, QByteArray("\220\037\240")) });
|
|
testTokens("<>", { Token(Type::String, QByteArray("")) });
|
|
|
|
testTokens("(Simple string)(Simple string)", { Token(Type::String, QByteArray("Simple string")), Token(Type::String, QByteArray("Simple string")) });
|
|
testTokens("(String with (brackets))(String with (brackets))", { Token(Type::String, QByteArray("String with (brackets)")), Token(Type::String, QByteArray("String with (brackets)")) });
|
|
testTokens("(String with \\( unbalanced brackets \\(\\))(String with \\( unbalanced brackets \\(\\))", { Token(Type::String, QByteArray("String with ( unbalanced brackets ()")), Token(Type::String, QByteArray("String with ( unbalanced brackets ()")) });
|
|
testTokens("()()", { Token(Type::String, QByteArray("")), Token(Type::String, QByteArray("")) });
|
|
testTokens("(Text with special character: \\n)(Text with special character: \\n)", { Token(Type::String, QByteArray("Text with special character: \n")), Token(Type::String, QByteArray("Text with special character: \n")) });
|
|
testTokens("(Text with special character: \\r)(Text with special character: \\r)", { Token(Type::String, QByteArray("Text with special character: \r")), Token(Type::String, QByteArray("Text with special character: \r")) });
|
|
testTokens("(Text with special character: \\t)(Text with special character: \\t)", { Token(Type::String, QByteArray("Text with special character: \t")), Token(Type::String, QByteArray("Text with special character: \t")) });
|
|
testTokens("(Text with special character: \\b)(Text with special character: \\b)", { Token(Type::String, QByteArray("Text with special character: \b")), Token(Type::String, QByteArray("Text with special character: \b")) });
|
|
testTokens("(Text with special character: \\f)(Text with special character: \\f)", { Token(Type::String, QByteArray("Text with special character: \f")), Token(Type::String, QByteArray("Text with special character: \f")) });
|
|
testTokens("(Text with special character: \\()(Text with special character: \\()", { Token(Type::String, QByteArray("Text with special character: (")), Token(Type::String, QByteArray("Text with special character: (")) });
|
|
testTokens("(Text with special character: \\))(Text with special character: \\))", { Token(Type::String, QByteArray("Text with special character: )")), Token(Type::String, QByteArray("Text with special character: )")) });
|
|
testTokens("(Text with special character: \\\\)(Text with special character: \\\\)", { Token(Type::String, QByteArray("Text with special character: \\")), Token(Type::String, QByteArray("Text with special character: \\")) });
|
|
testTokens("(\53)(\53)", { Token(Type::String, QByteArray("+")), Token(Type::String, QByteArray("+")) });
|
|
testTokens("(\0533)(\0533)", { Token(Type::String, QByteArray("+3")), Token(Type::String, QByteArray("+3")) });
|
|
testTokens("(\053)(\053)", { Token(Type::String, QByteArray("+")), Token(Type::String, QByteArray("+")) });
|
|
testTokens("(\053053)(\053053)", { Token(Type::String, QByteArray("+053")), Token(Type::String, QByteArray("+053")) });
|
|
testTokens("(\5)(\5)", { Token(Type::String, QByteArray("\5")), Token(Type::String, QByteArray("\5")) });
|
|
testTokens("<901FA3><901FA3>", { Token(Type::String, QByteArray("\220\037\243")), Token(Type::String, QByteArray("\220\037\243")) });
|
|
testTokens("<901fa3><901fa3>", { Token(Type::String, QByteArray("\220\037\243")), Token(Type::String, QByteArray("\220\037\243")) });
|
|
testTokens("<901fa><901fa>", { Token(Type::String, QByteArray("\220\037\240")), Token(Type::String, QByteArray("\220\037\240")) });
|
|
testTokens("<901FA><901FA>", { Token(Type::String, QByteArray("\220\037\240")), Token(Type::String, QByteArray("\220\037\240")) });
|
|
testTokens("<><>", { Token(Type::String, QByteArray("")), Token(Type::String, QByteArray("")) });
|
|
}
|
|
|
|
void LexicalAnalyzerTest::test_name()
|
|
{
|
|
using Token = pdf::PDFLexicalAnalyzer::Token;
|
|
using Type = pdf::PDFLexicalAnalyzer::TokenType;
|
|
|
|
testTokens("/Name123", { Token(Type::Name, QByteArray("Name123")) });
|
|
testTokens("/VeryLongName", { Token(Type::Name, QByteArray("VeryLongName")) });
|
|
testTokens("/A;Name_With^Various***Characters", { Token(Type::Name, QByteArray("A;Name_With^Various***Characters")) });
|
|
testTokens("/1.2", { Token(Type::Name, QByteArray("1.2")) });
|
|
testTokens("/$$", { Token(Type::Name, QByteArray("$$")) });
|
|
testTokens("/@MatchedPattern", { Token(Type::Name, QByteArray("@MatchedPattern")) });
|
|
testTokens("/.undefined", { Token(Type::Name, QByteArray(".undefined")) });
|
|
testTokens("/The#20Major#20And#20The#20#23", { Token(Type::Name, QByteArray("The Major And The #")) });
|
|
testTokens("/A#42", { Token(Type::Name, QByteArray("AB")) });
|
|
testTokens("/#20", { Token(Type::Name, QByteArray(" ")) });
|
|
testTokens("/#23#20#23/AB", { Token(Type::Name, QByteArray("# #")), Token(Type::Name, QByteArray("AB")) });
|
|
|
|
testTokens("/Name123/Name123", { Token(Type::Name, QByteArray("Name123")), Token(Type::Name, QByteArray("Name123")) });
|
|
testTokens("/VeryLongName/VeryLongName", { Token(Type::Name, QByteArray("VeryLongName")), Token(Type::Name, QByteArray("VeryLongName")) });
|
|
testTokens("/A;Name_With^Various***Characters/A;Name_With^Various***Characters", { Token(Type::Name, QByteArray("A;Name_With^Various***Characters")), Token(Type::Name, QByteArray("A;Name_With^Various***Characters")) });
|
|
testTokens("/1.2/1.2", { Token(Type::Name, QByteArray("1.2")), Token(Type::Name, QByteArray("1.2")) });
|
|
testTokens("/$$/$$", { Token(Type::Name, QByteArray("$$")), Token(Type::Name, QByteArray("$$")) });
|
|
testTokens("/@MatchedPattern/@MatchedPattern", { Token(Type::Name, QByteArray("@MatchedPattern")), Token(Type::Name, QByteArray("@MatchedPattern")) });
|
|
testTokens("/.undefined/.undefined", { Token(Type::Name, QByteArray(".undefined")), Token(Type::Name, QByteArray(".undefined")) });
|
|
testTokens("/The#20Major#20And#20The#20#23/The#20Major#20And#20The#20#23", { Token(Type::Name, QByteArray("The Major And The #")), Token(Type::Name, QByteArray("The Major And The #")) });
|
|
testTokens("/A#42/A#42", { Token(Type::Name, QByteArray("AB")), Token(Type::Name, QByteArray("AB")) });
|
|
testTokens("/#20/#20", { Token(Type::Name, QByteArray(" ")), Token(Type::Name, QByteArray(" ")) });
|
|
testTokens("/#23#20#23/AB/#23#20#23/AB", { Token(Type::Name, QByteArray("# #")), Token(Type::Name, QByteArray("AB")), Token(Type::Name, QByteArray("# #")), Token(Type::Name, QByteArray("AB")) });
|
|
}
|
|
|
|
void LexicalAnalyzerTest::test_bool()
|
|
{
|
|
using Token = pdf::PDFLexicalAnalyzer::Token;
|
|
using Type = pdf::PDFLexicalAnalyzer::TokenType;
|
|
|
|
testTokens("true", { Token(Type::Boolean, true) });
|
|
testTokens("false", { Token(Type::Boolean, false) });
|
|
testTokens("true false true false", { Token(Type::Boolean, true), Token(Type::Boolean, false), Token(Type::Boolean, true), Token(Type::Boolean, false) });
|
|
}
|
|
|
|
void LexicalAnalyzerTest::test_ad()
|
|
{
|
|
using Token = pdf::PDFLexicalAnalyzer::Token;
|
|
using Type = pdf::PDFLexicalAnalyzer::TokenType;
|
|
|
|
testTokens("<<", { Token(Type::DictionaryStart) });
|
|
testTokens("%comment\n<<", { Token(Type::DictionaryStart) });
|
|
testTokens(">>", { Token(Type::DictionaryEnd) });
|
|
testTokens("[", { Token(Type::ArrayStart) });
|
|
testTokens("]", { Token(Type::ArrayEnd) });
|
|
}
|
|
|
|
void LexicalAnalyzerTest::test_command()
|
|
{
|
|
using Token = pdf::PDFLexicalAnalyzer::Token;
|
|
using Type = pdf::PDFLexicalAnalyzer::TokenType;
|
|
|
|
testTokens("command", { Token(Type::Command, QByteArray("command")) });
|
|
testTokens("command1 command2", { Token(Type::Command, QByteArray("command1")), Token(Type::Command, QByteArray("command2")) });
|
|
}
|
|
|
|
void LexicalAnalyzerTest::test_invalid_input()
|
|
{
|
|
QByteArray bigNumber(500, '0');
|
|
bigNumber.front() = '1';
|
|
bigNumber.back() = 0;
|
|
|
|
QVERIFY_EXCEPTION_THROWN(scanWholeStream("(\\9adoctalnumber)"), pdf::PDFParserException);
|
|
QVERIFY_EXCEPTION_THROWN(scanWholeStream("(\\)"), pdf::PDFParserException);
|
|
QVERIFY_EXCEPTION_THROWN(scanWholeStream("123 456 +4-5"), pdf::PDFParserException);
|
|
QVERIFY_EXCEPTION_THROWN(scanWholeStream("123 456 +"), pdf::PDFParserException);
|
|
QVERIFY_EXCEPTION_THROWN(scanWholeStream("123 456 + 45"), pdf::PDFParserException);
|
|
QVERIFY_EXCEPTION_THROWN(scanWholeStream(bigNumber.constData()), pdf::PDFParserException);
|
|
QVERIFY_EXCEPTION_THROWN(scanWholeStream("/#Q1FF"), pdf::PDFParserException);
|
|
QVERIFY_EXCEPTION_THROWN(scanWholeStream("/#1QFF"), pdf::PDFParserException);
|
|
QVERIFY_EXCEPTION_THROWN(scanWholeStream("/# "), pdf::PDFParserException);
|
|
QVERIFY_EXCEPTION_THROWN(scanWholeStream("<A bad hexadecimal string>"), pdf::PDFParserException);
|
|
QVERIFY_EXCEPTION_THROWN(scanWholeStream("<1FA3"), pdf::PDFParserException);
|
|
QVERIFY_EXCEPTION_THROWN(scanWholeStream("<1FA"), pdf::PDFParserException);
|
|
QVERIFY_EXCEPTION_THROWN(scanWholeStream("> albatros"), pdf::PDFParserException);
|
|
QVERIFY_EXCEPTION_THROWN(scanWholeStream(")"), pdf::PDFParserException);
|
|
}
|
|
|
|
void LexicalAnalyzerTest::test_header_regexp()
|
|
{
|
|
std::regex regex(pdf::PDF_FILE_HEADER_REGEXP);
|
|
|
|
for (const char* string : { "%PDF-1.4", " %PDF-1.4abs", "%PDF-1.4", "%test %PDF %PDF-1.4", "%!PS-Adobe-3.0 PDF-1.4"})
|
|
{
|
|
std::cmatch cmatch;
|
|
const bool matched = std::regex_search(string, string + strlen(string), cmatch, regex);
|
|
QVERIFY(matched);
|
|
|
|
if (matched)
|
|
{
|
|
QVERIFY(cmatch.size() == 3);
|
|
QVERIFY(cmatch[1].matched || cmatch[2].matched);
|
|
}
|
|
}
|
|
}
|
|
|
|
void LexicalAnalyzerTest::test_flat_map()
|
|
{
|
|
using Map = pdf::PDFFlatMap<int, 2>;
|
|
|
|
struct Item
|
|
{
|
|
int order;
|
|
int number;
|
|
bool erase;
|
|
|
|
bool operator<(const Item& other) const { return order < other.order; }
|
|
};
|
|
|
|
for (int count = 1; count < 5; ++count)
|
|
{
|
|
std::vector<Item> items;
|
|
items.reserve(2 * count);
|
|
|
|
int order = 0;
|
|
for (int i = 0; i < count; ++i)
|
|
{
|
|
items.emplace_back(Item{order++, i, false});
|
|
items.emplace_back(Item{order++, i, true});
|
|
}
|
|
|
|
do
|
|
{
|
|
std::set<int> testSet;
|
|
Map testFlatMap;
|
|
|
|
for (const Item& item : items)
|
|
{
|
|
if (!item.erase)
|
|
{
|
|
testSet.insert(item.number);
|
|
testFlatMap.insert(item.number);
|
|
}
|
|
else
|
|
{
|
|
testSet.erase(item.number);
|
|
testFlatMap.erase(item.number);
|
|
}
|
|
|
|
QCOMPARE(testSet.size(), testFlatMap.size());
|
|
QCOMPARE(testSet.empty(), testFlatMap.empty());
|
|
|
|
for (const int testInteger : testSet)
|
|
{
|
|
QVERIFY(testFlatMap.search(testInteger));
|
|
}
|
|
}
|
|
|
|
} while (std::next_permutation(items.begin(), items.end()));
|
|
}
|
|
}
|
|
|
|
void LexicalAnalyzerTest::scanWholeStream(const char* stream)
|
|
{
|
|
pdf::PDFLexicalAnalyzer analyzer(stream, stream + strlen(stream));
|
|
|
|
// Scan whole stream
|
|
while (!analyzer.isAtEnd())
|
|
{
|
|
analyzer.fetch();
|
|
}
|
|
}
|
|
|
|
void LexicalAnalyzerTest::testTokens(const char* stream, const std::vector<pdf::PDFLexicalAnalyzer::Token>& tokens)
|
|
{
|
|
pdf::PDFLexicalAnalyzer analyzer(stream, stream + strlen(stream));
|
|
|
|
std::vector<pdf::PDFLexicalAnalyzer::Token> scanned;
|
|
scanned.reserve(tokens.size());
|
|
|
|
// Scan whole stream
|
|
while (!analyzer.isAtEnd())
|
|
{
|
|
scanned.emplace_back(analyzer.fetch());
|
|
}
|
|
|
|
// Format error message
|
|
QString actual = getStringFromTokens(scanned);
|
|
QString expected = getStringFromTokens(tokens);
|
|
|
|
// Now, compare scanned tokens
|
|
QVERIFY2(scanned == tokens, qPrintable(QString("stream: %1, actual = %2, expected = %3").arg(QString(stream), actual, expected)));
|
|
}
|
|
|
|
QString LexicalAnalyzerTest::getStringFromTokens(const std::vector<pdf::PDFLexicalAnalyzer::Token>& tokens)
|
|
{
|
|
QStringList stringTokens;
|
|
|
|
QMetaEnum metaEnum = QMetaEnum::fromType<pdf::PDFLexicalAnalyzer::TokenType>();
|
|
Q_ASSERT(metaEnum.isValid());
|
|
|
|
for (const pdf::PDFLexicalAnalyzer::Token& token : tokens)
|
|
{
|
|
QString tokenTypeAsString = metaEnum.valueToKey(static_cast<int>(token.type));
|
|
|
|
if (!token.data.isValid())
|
|
{
|
|
stringTokens << tokenTypeAsString;
|
|
}
|
|
else
|
|
{
|
|
stringTokens << QString("%1(%2)").arg(tokenTypeAsString, token.data.toString());
|
|
}
|
|
}
|
|
|
|
return QString("{ %1 }").arg(stringTokens.join(", "));
|
|
}
|
|
|
|
QTEST_APPLESS_MAIN(LexicalAnalyzerTest)
|
|
|
|
#include "tst_lexicalanalyzertest.moc"
|