File: tokenstream_tests.cpp

package info (click to toggle)
yaramod 3.6.0-1.1
  • links: PTS, VCS
  • area: main
  • in suites: bookworm
  • size: 3,924 kB
  • sloc: cpp: 52,237; python: 2,921; ansic: 71; sh: 70; makefile: 41
file content (151 lines) | stat: -rw-r--r-- 4,838 bytes parent folder | download | duplicates (2)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
/**
* @file tests/tokenstream_tests.cpp
* @brief Tests for the YARA literal.
* @copyright (c) 2019 Avast Software, licensed under the MIT license
*/

#include <gtest/gtest.h>
#include <iostream>

#include "yaramod/types/token_stream.h"
#include "yaramod/types/meta.h"

using namespace ::testing;

namespace yaramod {
namespace tests {

class TokenStreamTests : public Test {};

TEST_F(TokenStreamTests,
BoolLiteral) {
   auto l = Literal(true);
   ASSERT_TRUE(l.isBool());
   ASSERT_TRUE(l.getBool());
   ASSERT_EQ(l.getText(), "true");
   l = Literal(false);
   ASSERT_TRUE(l.isBool());
   ASSERT_FALSE(l.getBool());
   ASSERT_EQ(l.getText(), "false");
   l = Literal(true, "Yes");
   ASSERT_TRUE(l.isBool());
   ASSERT_TRUE(l.getBool());
   ASSERT_EQ(l.getText(), "Yes");
}

TEST_F(TokenStreamTests,
TokenStreamFindSimple) {
   TokenStream ts;

   auto found = ts.find(TokenType::META_KEY);
   ASSERT_EQ(found, ts.end());
   TokenIt key = ts.emplace_back(TokenType::META_KEY, "author");
   found = ts.find(TokenType::META_KEY);
   ASSERT_EQ(found, key);
}

TEST_F(TokenStreamTests,
TokenStreamFind) {
   TokenStream ts;
   TokenIt c1 = ts.emplace_back(TokenType::COMMENT, "/*c1*/");
   ts.emplace_back(TokenType::COMMENT, "/*c2*/");
   TokenIt k1 = ts.emplace_back(TokenType::META_KEY, "k1");
   TokenIt c3 = ts.emplace_back(TokenType::COMMENT, "/*c3*/");
   ts.emplace_back(TokenType::META_KEY, "k2");
   TokenIt c4 = ts.emplace_back(TokenType::COMMENT, "/*c4*/");
   ts.emplace_back(TokenType::COMMENT, "/*c5*/");
   ts.emplace_back(TokenType::META_KEY, "k3");

   ASSERT_EQ(ts.find(COMMENT), c1);
   ASSERT_EQ(ts.find(COMMENT, k1), c3);
   ASSERT_EQ(ts.find(COMMENT, c3), c3);
   ASSERT_EQ(ts.find(META_VALUE),         ts.end());
   ASSERT_EQ(ts.find(META_VALUE, c1),     ts.end());
   ASSERT_EQ(ts.find(META_VALUE, c1, c4), c4      );
}

TEST_F(TokenStreamTests,
TokenStreamFindBackwardsSimple) {
   TokenStream ts;

   auto found = ts.findBackwards(TokenType::META_KEY);
   ASSERT_EQ(found, ts.begin());
   TokenIt key = ts.emplace_back(TokenType::META_KEY, "author");
   found = ts.findBackwards(TokenType::META_KEY);
   ASSERT_EQ(found, key);
}

TEST_F(TokenStreamTests,
TokenStreamFindBackwards) {
   TokenStream ts;
   TokenIt c1 = ts.emplace_back(TokenType::COMMENT, "/*c1*/");
   TokenIt c2 = ts.emplace_back(TokenType::COMMENT, "/*c2*/");
   TokenIt k1 = ts.emplace_back(TokenType::META_KEY, "k1");
   TokenIt c3 = ts.emplace_back(TokenType::COMMENT, "/*c3*/");
   TokenIt k2 = ts.emplace_back(TokenType::META_KEY, "k2");
   TokenIt c4 = ts.emplace_back(TokenType::COMMENT, "/*c4*/");
   TokenIt c5 = ts.emplace_back(TokenType::COMMENT, "/*c5*/");
   ts.emplace_back(TokenType::META_KEY, "k3");

   ASSERT_EQ(ts.findBackwards(COMMENT), c5);
   ASSERT_EQ(ts.findBackwards(COMMENT, k1), c2);
   ASSERT_EQ(ts.findBackwards(COMMENT, c3), c2);
   ASSERT_EQ(ts.findBackwards(META_VALUE),         ts.end());
   ASSERT_EQ(ts.findBackwards(META_VALUE, c1),     c1);
   ASSERT_EQ(ts.findBackwards(META_KEY, c1, c4), k2      );
}

TEST_F(TokenStreamTests,
TokenStreamEmplaceBack) {
   TokenStream ts;

   TokenIt key = ts.emplace_back(TokenType::META_KEY, "author");
   ts.emplace_back(TokenType::EQ, "=");
   TokenIt value = ts.emplace_back(TokenType::META_VALUE, "Mr. Avastian");
   ASSERT_EQ(key->getPureText(), "author");
   ASSERT_EQ(value->getPureText(), "Mr. Avastian");
}

TEST_F(TokenStreamTests,
TokenStreamEmplace) {
   TokenStream ts;

   auto key = ts.emplace_back(TokenType::META_KEY, "author");
   ts.emplace_back(TokenType::EQ, "=");
   auto value = ts.emplace_back(TokenType::META_VALUE, "Mr. Avastian");
   auto comment = ts.emplace(value, TokenType::COMMENT, "/*comment about the author*/");
   ASSERT_EQ(key->getPureText(), "author");
   ASSERT_EQ(value->getPureText(), "Mr. Avastian");
   ASSERT_EQ(comment->getPureText(), "/*comment about the author*/");
}

TEST_F(TokenStreamTests,
TokenStreamPushBack) {
   TokenStream ts;
   Token t(TokenType::RULE_NAME, Literal("rule_name"));
   TokenIt name = ts.push_back(t);
   ASSERT_EQ(name->getPureText(), "rule_name");
}

TEST_F(TokenStreamTests,
TokenStreamErase) {
   TokenStream ts;
   TokenIt key = ts.emplace_back(TokenType::META_KEY, "author");
   ts.emplace_back(TokenType::META_KEY, "=");
   TokenIt comment1 = ts.emplace_back(TokenType::COMMENT, "/*comment before author name*/");
   TokenIt value = ts.emplace_back(TokenType::META_KEY, "author_name");
   TokenIt comment2 = ts.emplace_back(TokenType::COMMENT, "/*comment after author name*/");

   TokenIt behindErased = ts.erase(comment1);
   ASSERT_EQ(behindErased, value);
   behindErased = ts.erase(key, comment2);
   ASSERT_EQ(behindErased, comment2);
   ASSERT_EQ(comment2->getPureText(), "/*comment after author name*/");

   ts.clear();
   ASSERT_EQ(ts.size(), 0);
   ASSERT_TRUE(ts.empty());
}

}
}