@@ -63,6 +63,12 @@ gem 'haversine' |
||
63 | 63 |
gem 'omniauth-evernote' |
64 | 64 |
gem 'evernote_oauth' |
65 | 65 |
|
66 |
+# LocalFileAgent (watch functionality) |
|
67 |
+gem 'listen', '~> 3.0.5', require: false |
|
68 |
+ |
|
69 |
+# S3Agent |
|
70 |
+gem 'aws-sdk-core', '~> 2.2.15' |
|
71 |
+ |
|
66 | 72 |
# Optional Services. |
67 | 73 |
gem 'omniauth-37signals' # BasecampAgent |
68 | 74 |
gem 'omniauth-wunderlist', github: 'wunderlist/omniauth-wunderlist', ref: 'd0910d0396107b9302aa1bc50e74bb140990ccb8' |
@@ -75,7 +81,6 @@ unless Gem::Version.new(Bundler::VERSION) >= Gem::Version.new('1.5.0') |
||
75 | 81 |
end |
76 | 82 |
|
77 | 83 |
gem 'protected_attributes', '~>1.0.8' # This must be loaded before some other gems, like delayed_job. |
78 |
- |
|
79 | 84 |
gem 'ace-rails-ap', '~> 2.0.1' |
80 | 85 |
gem 'bootstrap-kaminari-views', '~> 0.0.3' |
81 | 86 |
gem 'bundler', '>= 1.5.0' |
@@ -110,6 +110,8 @@ GEM |
||
110 | 110 |
addressable (>= 2.3.1) |
111 | 111 |
extlib (>= 0.9.15) |
112 | 112 |
multi_json (>= 1.0.0) |
113 |
+ aws-sdk-core (2.2.15) |
|
114 |
+ jmespath (~> 1.0) |
|
113 | 115 |
bcrypt (3.1.10) |
114 | 116 |
better_errors (1.1.0) |
115 | 117 |
coderay (>= 1.0.0) |
@@ -279,6 +281,7 @@ GEM |
||
279 | 281 |
hypdf (1.0.7) |
280 | 282 |
httmultiparty (= 0.3.10) |
281 | 283 |
i18n (0.7.0) |
284 |
+ jmespath (1.1.3) |
|
282 | 285 |
jquery-rails (3.1.3) |
283 | 286 |
railties (>= 3.0, < 5.0) |
284 | 287 |
thor (>= 0.14, < 2.0) |
@@ -577,6 +580,7 @@ PLATFORMS |
||
577 | 580 |
|
578 | 581 |
DEPENDENCIES |
579 | 582 |
ace-rails-ap (~> 2.0.1) |
583 |
+ aws-sdk-core (~> 2.2.15) |
|
580 | 584 |
better_errors (~> 1.1) |
581 | 585 |
binding_of_caller |
582 | 586 |
bootstrap-kaminari-views (~> 0.0.3) |
@@ -621,6 +625,7 @@ DEPENDENCIES |
||
621 | 625 |
kramdown (~> 1.3.3) |
622 | 626 |
letter_opener_web |
623 | 627 |
liquid (~> 3.0.3) |
628 |
+ listen (~> 3.0.5) |
|
624 | 629 |
mini_magick |
625 | 630 |
mqtt |
626 | 631 |
multi_xml |
@@ -0,0 +1,58 @@ |
||
1 |
+module FileHandling |
|
2 |
+ extend ActiveSupport::Concern |
|
3 |
+ |
|
4 |
+ def get_file_pointer(file) |
|
5 |
+ { file_pointer: { file: file, agent_id: id } } |
|
6 |
+ end |
|
7 |
+ |
|
8 |
+ def get_io(event) |
|
9 |
+ return nil unless event.payload['file_pointer'] && |
|
10 |
+ event.payload['file_pointer']['file'] && |
|
11 |
+ event.payload['file_pointer']['agent_id'] |
|
12 |
+ event.user.agents.find(event.payload['file_pointer']['agent_id']).get_io(event.payload['file_pointer']['file']) |
|
13 |
+ end |
|
14 |
+ |
|
15 |
+ def emitting_file_handling_agent_description |
|
16 |
+ @emitting_file_handling_agent_description ||= |
|
17 |
+ "This agent only emits a 'file pointer', not the data inside the files, the following agents can consume the created events: `#{receiving_file_handling_agents.join('`, `')}`. Read more about the concept in the [wiki](https://github.com/cantino/huginn/wiki/How-Huginn-works-with-files)." |
|
18 |
+ end |
|
19 |
+ |
|
20 |
+ def receiving_file_handling_agent_description |
|
21 |
+ @receiving_file_handling_agent_description ||= |
|
22 |
+ "This agent can consume a 'file pointer' event from the following agents with no additional configuration: `#{emitting_file_handling_agents.join('`, `')}`. Read more about the concept in the [wiki](https://github.com/cantino/huginn/wiki/How-Huginn-works-with-files)." |
|
23 |
+ end |
|
24 |
+ |
|
25 |
+ private |
|
26 |
+ |
|
27 |
+ def emitting_file_handling_agents |
|
28 |
+ emitting_file_handling_agents = file_handling_agents.select { |a| a.emits_file_pointer? } |
|
29 |
+ emitting_file_handling_agents.map { |a| a.to_s.demodulize } |
|
30 |
+ end |
|
31 |
+ |
|
32 |
+ def receiving_file_handling_agents |
|
33 |
+ receiving_file_handling_agents = file_handling_agents.select { |a| a.consumes_file_pointer? } |
|
34 |
+ receiving_file_handling_agents.map { |a| a.to_s.demodulize } |
|
35 |
+ end |
|
36 |
+ |
|
37 |
+ def file_handling_agents |
|
38 |
+ @file_handling_agents ||= Agent.types.select{ |c| c.included_modules.include?(FileHandling) }.map { |d| d.name.constantize } |
|
39 |
+ end |
|
40 |
+ |
|
41 |
+ module ClassMethods |
|
42 |
+ def emits_file_pointer! |
|
43 |
+ @emits_file_pointer = true |
|
44 |
+ end |
|
45 |
+ |
|
46 |
+ def emits_file_pointer? |
|
47 |
+ !!@emits_file_pointer |
|
48 |
+ end |
|
49 |
+ |
|
50 |
+ def consumes_file_pointer! |
|
51 |
+ @consumes_file_pointer = true |
|
52 |
+ end |
|
53 |
+ |
|
54 |
+ def consumes_file_pointer? |
|
55 |
+ !!@consumes_file_pointer |
|
56 |
+ end |
|
57 |
+ end |
|
58 |
+end |
@@ -12,4 +12,8 @@ module WorkingHelpers |
||
12 | 12 |
def received_event_without_error? |
13 | 13 |
(last_receive_at.present? && last_error_log_at.blank?) || (last_receive_at.present? && last_error_log_at.present? && last_receive_at > last_error_log_at) |
14 | 14 |
end |
15 |
-end |
|
15 |
+ |
|
16 |
+ def checked_without_error? |
|
17 |
+ (last_check_at.present? && last_error_log_at.nil?) || (last_check_at.present? && last_error_log_at.present? && last_check_at > last_error_log_at) |
|
18 |
+ end |
|
19 |
+end |
@@ -0,0 +1,195 @@ |
||
1 |
+module Agents |
|
2 |
+ class CsvAgent < Agent |
|
3 |
+ include FormConfigurable |
|
4 |
+ include FileHandling |
|
5 |
+ |
|
6 |
+ cannot_be_scheduled! |
|
7 |
+ consumes_file_pointer! |
|
8 |
+ |
|
9 |
+ def default_options |
|
10 |
+ { |
|
11 |
+ 'mode' => 'parse', |
|
12 |
+ 'separator' => ',', |
|
13 |
+ 'use_fields' => '', |
|
14 |
+ 'output' => 'event_per_row', |
|
15 |
+ 'with_header' => 'true', |
|
16 |
+ 'data_path' => '$.data', |
|
17 |
+ 'data_key' => 'data' |
|
18 |
+ } |
|
19 |
+ end |
|
20 |
+ |
|
21 |
+ description do |
|
22 |
+ <<-MD |
|
23 |
+ The `CsvAgent` parses or serializes CSV data. When parsing, events can either be emitted for the entire CSV, or one per row. |
|
24 |
+ |
|
25 |
+ Set `mode` to `parse` to parse CSV from incoming event, when set to `serialize` the agent serilizes the data of events to CSV. |
|
26 |
+ |
|
27 |
+ ### Universal options |
|
28 |
+ |
|
29 |
+ Specify the `separator` which is used to seperate the fields from each other (default is `,`). |
|
30 |
+ |
|
31 |
+ `data_key` sets the key which contains the serialized CSV or parsed CSV data in emitted events. |
|
32 |
+ |
|
33 |
+ ### Parsing |
|
34 |
+ |
|
35 |
+ If `use_fields` is set to a comma seperated string and the CSV file contains field headers the agent will only extract the specified fields. |
|
36 |
+ |
|
37 |
+ `output` determines wheather one event per row is emitted or one event that includes all the rows. |
|
38 |
+ |
|
39 |
+ Set `with_header` to `true` if first line of the CSV file are field names. |
|
40 |
+ |
|
41 |
+ #{receiving_file_handling_agent_description} |
|
42 |
+ |
|
43 |
+ When receiving the CSV data in a regular event use [JSONPath](http://goessner.net/articles/JsonPath/) to select the path in `data_path`. `data_path` is only used when the received event does not contain a 'file pointer'. |
|
44 |
+ |
|
45 |
+ ### Serializing |
|
46 |
+ |
|
47 |
+ If `use_fields` is set to a comma seperated string and the first received event has a object at the specified `data_path` the generated CSV will only include the given fields. |
|
48 |
+ |
|
49 |
+ Set `with_header` to `true` to include a field header in the CSV. |
|
50 |
+ |
|
51 |
+ Use [JSONPath](http://goessner.net/articles/JsonPath/) in `data_path` to select with part of the received events should be serialized. |
|
52 |
+ MD |
|
53 |
+ end |
|
54 |
+ |
|
55 |
+ event_description do |
|
56 |
+ "Events will looks like this:\n\n %s" % if interpolated['mode'] == 'parse' |
|
57 |
+ rows = if boolify(interpolated['with_header']) |
|
58 |
+ [{'column' => 'row1 value1', 'column2' => 'row1 value2'}, {'column' => 'row2 value3', 'column2' => 'row2 value4'}] |
|
59 |
+ else |
|
60 |
+ [['row1 value1', 'row1 value2'], ['row2 value1', 'row2 value2']] |
|
61 |
+ end |
|
62 |
+ if interpolated['output'] == 'event_per_row' |
|
63 |
+ Utils.pretty_print(interpolated['data_key'] => rows[0]) |
|
64 |
+ else |
|
65 |
+ Utils.pretty_print(interpolated['data_key'] => rows) |
|
66 |
+ end |
|
67 |
+ else |
|
68 |
+ Utils.pretty_print(interpolated['data_key'] => '"generated","csv","data"' + "\n" + '"column1","column2","column3"') |
|
69 |
+ end |
|
70 |
+ end |
|
71 |
+ |
|
72 |
+ form_configurable :mode, type: :array, values: %w(parse serialize) |
|
73 |
+ form_configurable :separator, type: :string |
|
74 |
+ form_configurable :data_key, type: :string |
|
75 |
+ form_configurable :with_header, type: :boolean |
|
76 |
+ form_configurable :use_fields, type: :string |
|
77 |
+ form_configurable :output, type: :array, values: %w(event_per_row event_per_file) |
|
78 |
+ form_configurable :data_path, type: :string |
|
79 |
+ |
|
80 |
+ def validate_options |
|
81 |
+ if options['with_header'].blank? || ![true, false].include?(boolify(options['with_header'])) |
|
82 |
+ errors.add(:base, "The 'with_header' options is required and must be set to 'true' or 'false'") |
|
83 |
+ end |
|
84 |
+ if options['mode'] == 'serialize' && options['data_path'].blank? |
|
85 |
+ errors.add(:base, "When mode is set to serialize data_path has to be present.") |
|
86 |
+ end |
|
87 |
+ end |
|
88 |
+ |
|
89 |
+ def working? |
|
90 |
+ received_event_without_error? |
|
91 |
+ end |
|
92 |
+ |
|
93 |
+ def receive(incoming_events) |
|
94 |
+ case options['mode'] |
|
95 |
+ when 'parse' |
|
96 |
+ parse(incoming_events) |
|
97 |
+ when 'serialize' |
|
98 |
+ serialize(incoming_events) |
|
99 |
+ end |
|
100 |
+ end |
|
101 |
+ |
|
102 |
+ private |
|
103 |
+ def serialize(incoming_events) |
|
104 |
+ mo = interpolated(incoming_events.first) |
|
105 |
+ rows = rows_from_events(incoming_events, mo) |
|
106 |
+ csv = CSV.generate(col_sep: separator(mo), force_quotes: true ) do |csv| |
|
107 |
+ if boolify(mo['with_header']) && rows.first.is_a?(Hash) |
|
108 |
+ if mo['use_fields'].present? |
|
109 |
+ csv << extract_options(mo) |
|
110 |
+ else |
|
111 |
+ csv << rows.first.keys |
|
112 |
+ end |
|
113 |
+ end |
|
114 |
+ rows.each do |data| |
|
115 |
+ if data.is_a?(Hash) |
|
116 |
+ if mo['use_fields'].present? |
|
117 |
+ csv << data.extract!(*extract_options(mo)).values |
|
118 |
+ else |
|
119 |
+ csv << data.values |
|
120 |
+ end |
|
121 |
+ else |
|
122 |
+ csv << data |
|
123 |
+ end |
|
124 |
+ end |
|
125 |
+ end |
|
126 |
+ create_event payload: { mo['data_key'] => csv } |
|
127 |
+ end |
|
128 |
+ |
|
129 |
+ def rows_from_events(incoming_events, mo) |
|
130 |
+ [].tap do |rows| |
|
131 |
+ incoming_events.each do |event| |
|
132 |
+ data = Utils.value_at(event.payload, mo['data_path']) |
|
133 |
+ if data.is_a?(Array) && (data[0].is_a?(Array) || data[0].is_a?(Hash)) |
|
134 |
+ data.each { |row| rows << row } |
|
135 |
+ else |
|
136 |
+ rows << data |
|
137 |
+ end |
|
138 |
+ end |
|
139 |
+ end |
|
140 |
+ end |
|
141 |
+ |
|
142 |
+ def parse(incoming_events) |
|
143 |
+ incoming_events.each do |event| |
|
144 |
+ mo = interpolated(event) |
|
145 |
+ next unless io = local_get_io(event) |
|
146 |
+ if mo['output'] == 'event_per_row' |
|
147 |
+ parse_csv(io, mo) do |payload| |
|
148 |
+ create_event payload: { mo['data_key'] => payload } |
|
149 |
+ end |
|
150 |
+ else |
|
151 |
+ create_event payload: { mo['data_key'] => parse_csv(io, mo, []) } |
|
152 |
+ end |
|
153 |
+ end |
|
154 |
+ end |
|
155 |
+ |
|
156 |
+ def local_get_io(event) |
|
157 |
+ if io = get_io(event) |
|
158 |
+ io |
|
159 |
+ else |
|
160 |
+ Utils.value_at(event.payload, interpolated['data_path']) |
|
161 |
+ end |
|
162 |
+ end |
|
163 |
+ |
|
164 |
+ def parse_csv(io, mo, array = nil) |
|
165 |
+ CSV.new(io, col_sep: separator(mo), headers: boolify(mo['with_header'])).each do |row| |
|
166 |
+ if block_given? |
|
167 |
+ yield get_payload(row, mo) |
|
168 |
+ else |
|
169 |
+ array << get_payload(row, mo) |
|
170 |
+ end |
|
171 |
+ end |
|
172 |
+ array |
|
173 |
+ end |
|
174 |
+ |
|
175 |
+ def separator(mo) |
|
176 |
+ mo['separator'] == '\\t' ? "\t" : mo['separator'] |
|
177 |
+ end |
|
178 |
+ |
|
179 |
+ def get_payload(row, mo) |
|
180 |
+ if boolify(mo['with_header']) |
|
181 |
+ if mo['use_fields'].present? |
|
182 |
+ row.to_hash.extract!(*extract_options(mo)) |
|
183 |
+ else |
|
184 |
+ row.to_hash |
|
185 |
+ end |
|
186 |
+ else |
|
187 |
+ row |
|
188 |
+ end |
|
189 |
+ end |
|
190 |
+ |
|
191 |
+ def extract_options(mo) |
|
192 |
+ mo['use_fields'].split(',').map(&:strip) |
|
193 |
+ end |
|
194 |
+ end |
|
195 |
+end |
@@ -3,23 +3,42 @@ require 'time' |
||
3 | 3 |
|
4 | 4 |
module Agents |
5 | 5 |
class FtpsiteAgent < Agent |
6 |
- cannot_receive_events! |
|
6 |
+ include FileHandling |
|
7 | 7 |
default_schedule "every_12h" |
8 | 8 |
|
9 | 9 |
gem_dependency_check { defined?(Net::FTP) && defined?(Net::FTP::List) } |
10 | 10 |
|
11 |
- description <<-MD |
|
12 |
- The FTP Site Agent checks an FTP site and creates Events based on newly uploaded files in a directory. |
|
11 |
+ emits_file_pointer! |
|
13 | 12 |
|
14 |
- #{'## Include `net-ftp-list` in your Gemfile to use this Agent!' if dependencies_missing?} |
|
13 |
+ description do |
|
14 |
+ <<-MD |
|
15 |
+ The Ftp Site Agent checks an FTP site and creates Events based on newly uploaded files in a directory. When receiving events it creates files on the configured FTP server. |
|
15 | 16 |
|
17 |
+ #{'## Include `net-ftp-list` in your Gemfile to use this Agent!' if dependencies_missing?} |
|
16 | 18 |
|
17 |
- Specify a `url` that represents a directory of an FTP site to watch, and a list of `patterns` to match against file names. |
|
19 |
+ `mode` must be present and either `read` or `write`, in `read` mode the agent checks the FTP site for changed files, with `write` it writes received events to a file on the server. |
|
18 | 20 |
|
19 |
- Login credentials can be included in `url` if authentication is required. |
|
21 |
+ ### Universal options |
|
20 | 22 |
|
21 |
- Only files with a last modification time later than the `after` value, if specifed, are notified. |
|
22 |
- MD |
|
23 |
+ Specify a `url` that represents a directory of an FTP site to watch, and a list of `patterns` to match against file names. |
|
24 |
+ |
|
25 |
+ Login credentials can be included in `url` if authentication is required: `ftp://username:password@ftp.example.com/path`. Liquid formatting is supported as well: `ftp://{% credential ftp_credentials %}@ftp.example.com/` |
|
26 |
+ |
|
27 |
+ Optionally specify the encoding of the files you want to read/write in `force_encoding`, by default UTF-8 is used. |
|
28 |
+ |
|
29 |
+ ### Reading |
|
30 |
+ |
|
31 |
+ Only files with a last modification time later than the `after` value, if specifed, are emitted as event. |
|
32 |
+ |
|
33 |
+ ### Writing |
|
34 |
+ |
|
35 |
+ Specify the filename to use in `filename`, Liquid interpolation is possible to change the name per event. |
|
36 |
+ |
|
37 |
+ Use [Liquid](https://github.com/cantino/huginn/wiki/Formatting-Events-using-Liquid) templating in `data` to specify which part of the received event should be written. |
|
38 |
+ |
|
39 |
+ #{emitting_file_handling_agent_description} |
|
40 |
+ MD |
|
41 |
+ end |
|
23 | 42 |
|
24 | 43 |
event_description <<-MD |
25 | 44 |
Events look like this: |
@@ -32,42 +51,67 @@ module Agents |
||
32 | 51 |
MD |
33 | 52 |
|
34 | 53 |
def working? |
35 |
- event_created_within?(interpolated['expected_update_period_in_days']) && !recent_error_logs? |
|
54 |
+ if interpolated['mode'] == 'read' |
|
55 |
+ event_created_within?(interpolated['expected_update_period_in_days']) && !recent_error_logs? |
|
56 |
+ else |
|
57 |
+ received_event_without_error? |
|
58 |
+ end |
|
36 | 59 |
end |
37 | 60 |
|
38 | 61 |
def default_options |
39 | 62 |
{ |
63 |
+ 'mode' => 'read', |
|
40 | 64 |
'expected_update_period_in_days' => "1", |
41 | 65 |
'url' => "ftp://example.org/pub/releases/", |
42 | 66 |
'patterns' => [ |
43 | 67 |
'foo-*.tar.gz', |
44 | 68 |
], |
45 | 69 |
'after' => Time.now.iso8601, |
70 |
+ 'force_encoding' => '', |
|
71 |
+ 'filename' => '', |
|
72 |
+ 'data' => '{{ data }}' |
|
46 | 73 |
} |
47 | 74 |
end |
48 | 75 |
|
49 | 76 |
def validate_options |
50 | 77 |
# Check for required fields |
51 | 78 |
begin |
52 |
- url = options['url'] |
|
53 |
- String === url or raise |
|
54 |
- uri = URI(url) |
|
55 |
- URI::FTP === uri or raise |
|
56 |
- errors.add(:base, "url must end with a slash") unless uri.path.end_with?('/') |
|
79 |
+ if !options['url'].include?('{{') |
|
80 |
+ url = interpolated['url'] |
|
81 |
+ String === url or raise |
|
82 |
+ uri = URI(url) |
|
83 |
+ URI::FTP === uri or raise |
|
84 |
+ errors.add(:base, "url must end with a slash") if uri.path.present? && !uri.path.end_with?('/') |
|
85 |
+ end |
|
57 | 86 |
rescue |
58 | 87 |
errors.add(:base, "url must be a valid FTP URL") |
59 | 88 |
end |
60 | 89 |
|
61 |
- patterns = options['patterns'] |
|
62 |
- case patterns |
|
63 |
- when Array |
|
64 |
- if patterns.empty? |
|
65 |
- errors.add(:base, "patterns must not be empty") |
|
90 |
+ options['mode'] = 'read' if options['mode'].blank? && new_record? |
|
91 |
+ if options['mode'].blank? || !['read', 'write'].include?(options['mode']) |
|
92 |
+ errors.add(:base, "The 'mode' option is required and must be set to 'read' or 'write'") |
|
93 |
+ end |
|
94 |
+ |
|
95 |
+ case interpolated['mode'] |
|
96 |
+ when 'read' |
|
97 |
+ patterns = options['patterns'] |
|
98 |
+ case patterns |
|
99 |
+ when Array |
|
100 |
+ if patterns.empty? |
|
101 |
+ errors.add(:base, "patterns must not be empty") |
|
102 |
+ end |
|
103 |
+ when nil, '' |
|
104 |
+ errors.add(:base, "patterns must be specified") |
|
105 |
+ else |
|
106 |
+ errors.add(:base, "patterns must be an array") |
|
107 |
+ end |
|
108 |
+ when 'write' |
|
109 |
+ if options['filename'].blank? |
|
110 |
+ errors.add(:base, "filename must be specified in 'write' mode") |
|
111 |
+ end |
|
112 |
+ if options['data'].blank? |
|
113 |
+ errors.add(:base, "data must be specified in 'write' mode") |
|
66 | 114 |
end |
67 |
- when nil, '' |
|
68 |
- errors.add(:base, "patterns must be specified") |
|
69 |
- else |
|
70 |
- errors.add(:base, "patterns must be an array") |
|
71 | 115 |
end |
72 | 116 |
|
73 | 117 |
# Check for optional fields |
@@ -85,6 +129,7 @@ module Agents |
||
85 | 129 |
end |
86 | 130 |
|
87 | 131 |
def check |
132 |
+ return if interpolated['mode'] != 'read' |
|
88 | 133 |
saving_entries do |found| |
89 | 134 |
each_entry { |filename, mtime| |
90 | 135 |
found[filename, mtime] |
@@ -92,6 +137,17 @@ module Agents |
||
92 | 137 |
end |
93 | 138 |
end |
94 | 139 |
|
140 |
+ def receive(incoming_events) |
|
141 |
+ return if interpolated['mode'] != 'write' |
|
142 |
+ incoming_events.each do |event| |
|
143 |
+ mo = interpolated(event) |
|
144 |
+ mo['data'].encode!(interpolated['force_encoding'], invalid: :replace, undef: :replace) if interpolated['force_encoding'].present? |
|
145 |
+ open_ftp(base_uri) do |ftp| |
|
146 |
+ ftp.storbinary("STOR #{mo['filename']}", StringIO.new(mo['data']), Net::FTP::DEFAULT_BLOCKSIZE) |
|
147 |
+ end |
|
148 |
+ end |
|
149 |
+ end |
|
150 |
+ |
|
95 | 151 |
def each_entry |
96 | 152 |
patterns = interpolated['patterns'] |
97 | 153 |
|
@@ -147,9 +203,10 @@ module Agents |
||
147 | 203 |
|
148 | 204 |
ftp.passive = true |
149 | 205 |
|
150 |
- path = uri.path.chomp('/') |
|
151 |
- log "Changing directory to #{path}" |
|
152 |
- ftp.chdir(path) |
|
206 |
+ if (path = uri.path.chomp('/')).present? |
|
207 |
+ log "Changing directory to #{path}" |
|
208 |
+ ftp.chdir(path) |
|
209 |
+ end |
|
153 | 210 |
|
154 | 211 |
yield ftp |
155 | 212 |
ensure |
@@ -176,17 +233,28 @@ module Agents |
||
176 | 233 |
new_files.sort_by { |filename| |
177 | 234 |
found_entries[filename] |
178 | 235 |
}.each { |filename| |
179 |
- create_event payload: { |
|
236 |
+ create_event payload: get_file_pointer(filename).merge({ |
|
180 | 237 |
'url' => (base_uri + uri_path_escape(filename)).to_s, |
181 | 238 |
'filename' => filename, |
182 | 239 |
'timestamp' => found_entries[filename], |
183 |
- } |
|
240 |
+ }) |
|
184 | 241 |
} |
185 | 242 |
|
186 | 243 |
memory['known_entries'] = found_entries |
187 | 244 |
save! |
188 | 245 |
end |
189 | 246 |
|
247 |
+ def get_io(file) |
|
248 |
+ data = StringIO.new |
|
249 |
+ open_ftp(base_uri) do |ftp| |
|
250 |
+ ftp.getbinaryfile(file, nil) do |chunk| |
|
251 |
+ data.write chunk.force_encoding(options['force_encoding'].presence || 'UTF-8') |
|
252 |
+ end |
|
253 |
+ end |
|
254 |
+ data.rewind |
|
255 |
+ data |
|
256 |
+ end |
|
257 |
+ |
|
190 | 258 |
private |
191 | 259 |
|
192 | 260 |
def is_positive_integer?(value) |
@@ -0,0 +1,190 @@ |
||
1 |
+module Agents |
|
2 |
+ class LocalFileAgent < Agent |
|
3 |
+ include LongRunnable |
|
4 |
+ include FormConfigurable |
|
5 |
+ include FileHandling |
|
6 |
+ |
|
7 |
+ emits_file_pointer! |
|
8 |
+ |
|
9 |
+ default_schedule 'every_1h' |
|
10 |
+ |
|
11 |
+ def self.should_run? |
|
12 |
+ ENV['ENABLE_INSECURE_AGENTS'] == "true" |
|
13 |
+ end |
|
14 |
+ |
|
15 |
+ description do |
|
16 |
+ <<-MD |
|
17 |
+ The LocalFileAgent can watch a file/directory for changes or emit an event for every file in that directory. When receiving an event it writes the received data into a file. |
|
18 |
+ |
|
19 |
+ `mode` determines if the agent is emitting events for (changed) files or writing received event data to disk. |
|
20 |
+ |
|
21 |
+ ### Reading |
|
22 |
+ |
|
23 |
+ When `watch` is set to `true` the LocalFileAgent will watch the specified `path` for changes, the schedule is ignored and the file system is watched continuously. An event will be emitted for every detected change. |
|
24 |
+ |
|
25 |
+ When `watch` is set to `false` the agent will emit an event for every file in the directory on each scheduled run. |
|
26 |
+ |
|
27 |
+ #{emitting_file_handling_agent_description} |
|
28 |
+ |
|
29 |
+ ### Writing |
|
30 |
+ |
|
31 |
+ Every event will be writting into a file at `path`, Liquid interpolation is possible to change the path per event. |
|
32 |
+ |
|
33 |
+ When `append` is true the received data will be appended to the file. |
|
34 |
+ |
|
35 |
+ Use [Liquid](https://github.com/cantino/huginn/wiki/Formatting-Events-using-Liquid) templating in `data` to specify which part of the received event should be written. |
|
36 |
+ |
|
37 |
+ *Warning*: This type of Agent can read and write any file the user that runs the Huginn server has access to, and is #{Agents::LocalFileAgent.should_run? ? "**currently enabled**" : "**currently disabled**"}. |
|
38 |
+ Only enable this Agent if you trust everyone using your Huginn installation. |
|
39 |
+ You can enable this Agent in your .env file by setting `ENABLE_INSECURE_AGENTS` to `true`. |
|
40 |
+ MD |
|
41 |
+ end |
|
42 |
+ |
|
43 |
+ event_description do |
|
44 |
+ "Events will looks like this:\n\n %s" % if boolify(interpolated['watch']) |
|
45 |
+ Utils.pretty_print( |
|
46 |
+ "file_pointer" => { |
|
47 |
+ "file" => "/tmp/test/filename", |
|
48 |
+ "agent_id" => id |
|
49 |
+ }, |
|
50 |
+ "event_type" => "modified/added/removed" |
|
51 |
+ ) |
|
52 |
+ else |
|
53 |
+ Utils.pretty_print( |
|
54 |
+ "file_pointer" => { |
|
55 |
+ "file" => "/tmp/test/filename", |
|
56 |
+ "agent_id" => id |
|
57 |
+ } |
|
58 |
+ ) |
|
59 |
+ end |
|
60 |
+ end |
|
61 |
+ |
|
62 |
+ def default_options |
|
63 |
+ { |
|
64 |
+ 'mode' => 'read', |
|
65 |
+ 'watch' => 'true', |
|
66 |
+ 'append' => 'false', |
|
67 |
+ 'path' => "", |
|
68 |
+ 'data' => '{{ data }}' |
|
69 |
+ } |
|
70 |
+ end |
|
71 |
+ |
|
72 |
+ form_configurable :mode, type: :array, values: %w(read write) |
|
73 |
+ form_configurable :watch, type: :array, values: %w(true false) |
|
74 |
+ form_configurable :path, type: :string |
|
75 |
+ form_configurable :append, type: :boolean |
|
76 |
+ form_configurable :data, type: :string |
|
77 |
+ |
|
78 |
+ def validate_options |
|
79 |
+ if options['mode'].blank? || !['read', 'write'].include?(options['mode']) |
|
80 |
+ errors.add(:base, "The 'mode' option is required and must be set to 'read' or 'write'") |
|
81 |
+ end |
|
82 |
+ if options['watch'].blank? || ![true, false].include?(boolify(options['watch'])) |
|
83 |
+ errors.add(:base, "The 'watch' option is required and must be set to 'true' or 'false'") |
|
84 |
+ end |
|
85 |
+ if options['append'].blank? || ![true, false].include?(boolify(options['append'])) |
|
86 |
+ errors.add(:base, "The 'append' option is required and must be set to 'true' or 'false'") |
|
87 |
+ end |
|
88 |
+ if options['path'].blank? |
|
89 |
+ errors.add(:base, "The 'path' option is required.") |
|
90 |
+ end |
|
91 |
+ end |
|
92 |
+ |
|
93 |
+ def working? |
|
94 |
+ should_run?(false) && ((interpolated['mode'] == 'read' && check_path_existance && checked_without_error?) || |
|
95 |
+ (interpolated['mode'] == 'write' && received_event_without_error?)) |
|
96 |
+ end |
|
97 |
+ |
|
98 |
+ def check |
|
99 |
+ return if interpolated['mode'] != 'read' || boolify(interpolated['watch']) || !should_run? |
|
100 |
+ return unless check_path_existance(true) |
|
101 |
+ if File.directory?(expanded_path) |
|
102 |
+ Dir.glob(File.join(expanded_path, '*')).select { |f| File.file?(f) } |
|
103 |
+ else |
|
104 |
+ [expanded_path] |
|
105 |
+ end.each do |file| |
|
106 |
+ create_event payload: get_file_pointer(file) |
|
107 |
+ end |
|
108 |
+ end |
|
109 |
+ |
|
110 |
+ def receive(incoming_events) |
|
111 |
+ return if interpolated['mode'] != 'write' || !should_run? |
|
112 |
+ incoming_events.each do |event| |
|
113 |
+ mo = interpolated(event) |
|
114 |
+ File.open(File.expand_path(mo['path']), boolify(mo['append']) ? 'a' : 'w') do |file| |
|
115 |
+ file.write(mo['data']) |
|
116 |
+ end |
|
117 |
+ end |
|
118 |
+ end |
|
119 |
+ |
|
120 |
+ def start_worker? |
|
121 |
+ interpolated['mode'] == 'read' && boolify(interpolated['watch']) && should_run? && check_path_existance |
|
122 |
+ end |
|
123 |
+ |
|
124 |
+ def check_path_existance(log = true) |
|
125 |
+ if !File.exist?(expanded_path) |
|
126 |
+ error("File or directory '#{expanded_path}' does not exist") if log |
|
127 |
+ return false |
|
128 |
+ end |
|
129 |
+ true |
|
130 |
+ end |
|
131 |
+ |
|
132 |
+ def get_io(file) |
|
133 |
+ File.open(file, 'r') |
|
134 |
+ end |
|
135 |
+ |
|
136 |
+ def expanded_path |
|
137 |
+ @expanded_path ||= File.expand_path(interpolated['path']) |
|
138 |
+ end |
|
139 |
+ |
|
140 |
+ private |
|
141 |
+ |
|
142 |
+ def should_run?(log = true) |
|
143 |
+ if self.class.should_run? |
|
144 |
+ true |
|
145 |
+ else |
|
146 |
+ error("Unable to run because insecure agents are not enabled. Set ENABLE_INSECURE_AGENTS to true in the Huginn .env configuration.") if log |
|
147 |
+ false |
|
148 |
+ end |
|
149 |
+ end |
|
150 |
+ |
|
151 |
+ class Worker < LongRunnable::Worker |
|
152 |
+ def setup |
|
153 |
+ require 'listen' |
|
154 |
+ @listener = Listen.to(*listen_options, &method(:callback)) |
|
155 |
+ end |
|
156 |
+ |
|
157 |
+ def run |
|
158 |
+ sleep unless agent.check_path_existance(true) |
|
159 |
+ |
|
160 |
+ @listener.start |
|
161 |
+ sleep |
|
162 |
+ end |
|
163 |
+ |
|
164 |
+ def stop |
|
165 |
+ @listener.stop |
|
166 |
+ end |
|
167 |
+ |
|
168 |
+ private |
|
169 |
+ |
|
170 |
+ def callback(*changes) |
|
171 |
+ AgentRunner.with_connection do |
|
172 |
+ changes.zip([:modified, :added, :removed]).each do |files, event_type| |
|
173 |
+ files.each do |file| |
|
174 |
+ agent.create_event payload: agent.get_file_pointer(file).merge(event_type: event_type) |
|
175 |
+ end |
|
176 |
+ end |
|
177 |
+ agent.touch(:last_check_at) |
|
178 |
+ end |
|
179 |
+ end |
|
180 |
+ |
|
181 |
+ def listen_options |
|
182 |
+ if File.directory?(agent.expanded_path) |
|
183 |
+ [agent.expanded_path, ignore!: [] ] |
|
184 |
+ else |
|
185 |
+ [File.dirname(agent.expanded_path), { ignore!: [], only: /\A#{Regexp.escape(File.basename(agent.expanded_path))}\z/ } ] |
|
186 |
+ end |
|
187 |
+ end |
|
188 |
+ end |
|
189 |
+ end |
|
190 |
+end |
@@ -0,0 +1,50 @@ |
||
1 |
+module Agents |
|
2 |
+ class ReadFileAgent < Agent |
|
3 |
+ include FormConfigurable |
|
4 |
+ include FileHandling |
|
5 |
+ |
|
6 |
+ cannot_be_scheduled! |
|
7 |
+ consumes_file_pointer! |
|
8 |
+ |
|
9 |
+ def default_options |
|
10 |
+ { |
|
11 |
+ 'data_key' => 'data' |
|
12 |
+ } |
|
13 |
+ end |
|
14 |
+ |
|
15 |
+ description do |
|
16 |
+ <<-MD |
|
17 |
+ The ReadFileAgent takes events from `FileHandling` agents, reads the file, and emits the contents as a string. |
|
18 |
+ |
|
19 |
+ `data_key` specifies the key of the emitted event which contains the file contents. |
|
20 |
+ |
|
21 |
+ #{receiving_file_handling_agent_description} |
|
22 |
+ MD |
|
23 |
+ end |
|
24 |
+ |
|
25 |
+ event_description <<-MD |
|
26 |
+ { |
|
27 |
+ "data" => '...' |
|
28 |
+ } |
|
29 |
+ MD |
|
30 |
+ |
|
31 |
+ form_configurable :data_key, type: :string |
|
32 |
+ |
|
33 |
+ def validate_options |
|
34 |
+ if options['data_key'].blank? |
|
35 |
+ errors.add(:base, "The 'data_key' options is required.") |
|
36 |
+ end |
|
37 |
+ end |
|
38 |
+ |
|
39 |
+ def working? |
|
40 |
+ received_event_without_error? |
|
41 |
+ end |
|
42 |
+ |
|
43 |
+ def receive(incoming_events) |
|
44 |
+ incoming_events.each do |event| |
|
45 |
+ next unless io = get_io(event) |
|
46 |
+ create_event payload: { interpolated['data_key'] => io.read } |
|
47 |
+ end |
|
48 |
+ end |
|
49 |
+ end |
|
50 |
+end |
@@ -0,0 +1,206 @@ |
||
1 |
+module Agents |
|
2 |
+ class S3Agent < Agent |
|
3 |
+ include FormConfigurable |
|
4 |
+ include FileHandling |
|
5 |
+ |
|
6 |
+ emits_file_pointer! |
|
7 |
+ no_bulk_receive! |
|
8 |
+ |
|
9 |
+ default_schedule 'every_1h' |
|
10 |
+ |
|
11 |
+ gem_dependency_check { defined?(Aws::S3) } |
|
12 |
+ |
|
13 |
+ description do |
|
14 |
+ <<-MD |
|
15 |
+ The S3Agent can watch a bucket for changes or emit an event for every file in that bucket. When receiving events, it writes the data into a file on S3. |
|
16 |
+ |
|
17 |
+ #{'## Include `aws-sdk-core` in your Gemfile to use this Agent!' if dependencies_missing?} |
|
18 |
+ |
|
19 |
+ `mode` must be present and either `read` or `write`, in `read` mode the agent checks the S3 bucket for changed files, with `write` it writes received events to a file in the bucket. |
|
20 |
+ |
|
21 |
+ ### Universal options |
|
22 |
+ |
|
23 |
+ To use credentials for the `access_key` and `access_key_secret` use the liquid `credential` tag like so `{% credential name-of-credential %}` |
|
24 |
+ |
|
25 |
+ Select the `region` in which the bucket was created. |
|
26 |
+ |
|
27 |
+ ### Reading |
|
28 |
+ |
|
29 |
+ When `watch` is set to `true` the S3Agent will watch the specified `bucket` for changes. An event will be emitted for every detected change. |
|
30 |
+ |
|
31 |
+ When `watch` is set to `false` the agent will emit an event for every file in the bucket on each sheduled run. |
|
32 |
+ |
|
33 |
+ #{emitting_file_handling_agent_description} |
|
34 |
+ |
|
35 |
+ ### Writing |
|
36 |
+ |
|
37 |
+ Specify the filename to use in `filename`, Liquid interpolation is possible to change the name per event. |
|
38 |
+ |
|
39 |
+ Use [Liquid](https://github.com/cantino/huginn/wiki/Formatting-Events-using-Liquid) templating in `data` to specify which part of the received event should be written. |
|
40 |
+ MD |
|
41 |
+ end |
|
42 |
+ |
|
43 |
+ event_description do |
|
44 |
+ "Events will looks like this:\n\n %s" % if boolify(interpolated['watch']) |
|
45 |
+ Utils.pretty_print({ |
|
46 |
+ "file_pointer" => { |
|
47 |
+ "file" => "filename", |
|
48 |
+ "agent_id" => id |
|
49 |
+ }, |
|
50 |
+ "event_type" => "modified/added/removed" |
|
51 |
+ }) |
|
52 |
+ else |
|
53 |
+ Utils.pretty_print({ |
|
54 |
+ "file_pointer" => { |
|
55 |
+ "file" => "filename", |
|
56 |
+ "agent_id" => id |
|
57 |
+ } |
|
58 |
+ }) |
|
59 |
+ end |
|
60 |
+ end |
|
61 |
+ |
|
62 |
+ def default_options |
|
63 |
+ { |
|
64 |
+ 'mode' => 'read', |
|
65 |
+ 'access_key_id' => '', |
|
66 |
+ 'access_key_secret' => '', |
|
67 |
+ 'watch' => 'true', |
|
68 |
+ 'bucket' => "", |
|
69 |
+ 'data' => '{{ data }}' |
|
70 |
+ } |
|
71 |
+ end |
|
72 |
+ |
|
73 |
+ form_configurable :mode, type: :array, values: %w(read write) |
|
74 |
+ form_configurable :access_key_id, roles: :validatable |
|
75 |
+ form_configurable :access_key_secret, roles: :validatable |
|
76 |
+ form_configurable :region, type: :array, values: %w(us-east-1 us-west-1 us-west-2 eu-west-1 eu-central-1 ap-southeast-1 ap-southeast-2 ap-northeast-1 ap-northeast-2 sa-east-1) |
|
77 |
+ form_configurable :watch, type: :array, values: %w(true false) |
|
78 |
+ form_configurable :bucket, roles: :completable |
|
79 |
+ form_configurable :filename |
|
80 |
+ form_configurable :data |
|
81 |
+ |
|
82 |
+ def validate_options |
|
83 |
+ if options['mode'].blank? || !['read', 'write'].include?(options['mode']) |
|
84 |
+ errors.add(:base, "The 'mode' option is required and must be set to 'read' or 'write'") |
|
85 |
+ end |
|
86 |
+ if options['bucket'].blank? |
|
87 |
+ errors.add(:base, "The 'bucket' option is required.") |
|
88 |
+ end |
|
89 |
+ if options['region'].blank? |
|
90 |
+ errors.add(:base, "The 'region' option is required.") |
|
91 |
+ end |
|
92 |
+ |
|
93 |
+ case interpolated['mode'] |
|
94 |
+ when 'read' |
|
95 |
+ if options['watch'].blank? || ![true, false].include?(boolify(options['watch'])) |
|
96 |
+ errors.add(:base, "The 'watch' option is required and must be set to 'true' or 'false'") |
|
97 |
+ end |
|
98 |
+ when 'write' |
|
99 |
+ if options['filename'].blank? |
|
100 |
+ errors.add(:base, "filename must be specified in 'write' mode") |
|
101 |
+ end |
|
102 |
+ if options['data'].blank? |
|
103 |
+ errors.add(:base, "data must be specified in 'write' mode") |
|
104 |
+ end |
|
105 |
+ end |
|
106 |
+ end |
|
107 |
+ |
|
108 |
+ def validate_access_key_id |
|
109 |
+ !!buckets |
|
110 |
+ end |
|
111 |
+ |
|
112 |
+ def validate_access_key_secret |
|
113 |
+ !!buckets |
|
114 |
+ end |
|
115 |
+ |
|
116 |
+ def complete_bucket |
|
117 |
+ (buckets || []).collect { |room| {text: room.name, id: room.name} } |
|
118 |
+ end |
|
119 |
+ |
|
120 |
+ def working? |
|
121 |
+ checked_without_error? |
|
122 |
+ end |
|
123 |
+ |
|
124 |
+ def check |
|
125 |
+ return if interpolated['mode'] != 'read' |
|
126 |
+ contents = safely do |
|
127 |
+ get_bucket_contents |
|
128 |
+ end |
|
129 |
+ if boolify(interpolated['watch']) |
|
130 |
+ watch(contents) |
|
131 |
+ else |
|
132 |
+ contents.each do |key, _| |
|
133 |
+ create_event payload: get_file_pointer(key) |
|
134 |
+ end |
|
135 |
+ end |
|
136 |
+ end |
|
137 |
+ |
|
138 |
+ def get_io(file) |
|
139 |
+ client.get_object(bucket: interpolated['bucket'], key: file).body |
|
140 |
+ end |
|
141 |
+ |
|
142 |
+ def receive(incoming_events) |
|
143 |
+ return if interpolated['mode'] != 'write' |
|
144 |
+ incoming_events.each do |event| |
|
145 |
+ safely do |
|
146 |
+ mo = interpolated(event) |
|
147 |
+ client.put_object(bucket: mo['bucket'], key: mo['filename'], body: mo['data']) |
|
148 |
+ end |
|
149 |
+ end |
|
150 |
+ end |
|
151 |
+ |
|
152 |
+ private |
|
153 |
+ |
|
154 |
+ def safely |
|
155 |
+ yield |
|
156 |
+ rescue Aws::S3::Errors::AccessDenied => e |
|
157 |
+ error("Could not access '#{interpolated['bucket']}' #{e.class} #{e.message}") |
|
158 |
+ rescue Aws::S3::Errors::ServiceError =>e |
|
159 |
+ error("#{e.class}: #{e.message}") |
|
160 |
+ end |
|
161 |
+ |
|
162 |
+ def watch(contents) |
|
163 |
+ if last_check_at.nil? |
|
164 |
+ self.memory['seen_contents'] = contents |
|
165 |
+ return |
|
166 |
+ end |
|
167 |
+ |
|
168 |
+ new_memory = contents.dup |
|
169 |
+ |
|
170 |
+ memory['seen_contents'].each do |key, etag| |
|
171 |
+ if contents[key].blank? |
|
172 |
+ create_event payload: get_file_pointer(key).merge(event_type: :removed) |
|
173 |
+ elsif contents[key] != etag |
|
174 |
+ create_event payload: get_file_pointer(key).merge(event_type: :modified) |
|
175 |
+ end |
|
176 |
+ contents.delete(key) |
|
177 |
+ end |
|
178 |
+ contents.each do |key, etag| |
|
179 |
+ create_event payload: get_file_pointer(key).merge(event_type: :added) |
|
180 |
+ end |
|
181 |
+ |
|
182 |
+ self.memory['seen_contents'] = new_memory |
|
183 |
+ end |
|
184 |
+ |
|
185 |
+ def get_bucket_contents |
|
186 |
+ contents = {} |
|
187 |
+ client.list_objects(bucket: interpolated['bucket']).each do |response| |
|
188 |
+ response.contents.each do |file| |
|
189 |
+ contents[file.key] = file.etag |
|
190 |
+ end |
|
191 |
+ end |
|
192 |
+ contents |
|
193 |
+ end |
|
194 |
+ |
|
195 |
+ def client |
|
196 |
+ @client ||= Aws::S3::Client.new(credentials: Aws::Credentials.new(interpolated['access_key_id'], interpolated['access_key_secret']), |
|
197 |
+ region: interpolated['region']) |
|
198 |
+ end |
|
199 |
+ |
|
200 |
+ def buckets(log = false) |
|
201 |
+ @buckets ||= client.list_buckets.buckets |
|
202 |
+ rescue Aws::S3::Errors::ServiceError => e |
|
203 |
+ false |
|
204 |
+ end |
|
205 |
+ end |
|
206 |
+end |
@@ -0,0 +1,15 @@ |
||
1 |
+class AddModeOptionToFtpsiteAgents < ActiveRecord::Migration |
|
2 |
+ def up |
|
3 |
+ Agents::FtpsiteAgent.find_each do |agent| |
|
4 |
+ agent.options['mode'] = 'read' |
|
5 |
+ agent.save!(validate: false) |
|
6 |
+ end |
|
7 |
+ end |
|
8 |
+ |
|
9 |
+ def down |
|
10 |
+ Agents::FtpsiteAgent.find_each do |agent| |
|
11 |
+ agent.options.delete 'mode' |
|
12 |
+ agent.save!(validate: false) |
|
13 |
+ end |
|
14 |
+ end |
|
15 |
+end |
@@ -118,5 +118,6 @@ end |
||
118 | 118 |
|
119 | 119 |
require 'agents/twitter_stream_agent' |
120 | 120 |
require 'agents/jabber_agent' |
121 |
+require 'agents/local_file_agent' |
|
121 | 122 |
require 'huginn_scheduler' |
122 | 123 |
require 'delayed_job_worker' |
@@ -15,7 +15,7 @@ module Utils |
||
15 | 15 |
def self.pretty_print(struct, indent = true) |
16 | 16 |
output = JSON.pretty_generate(struct) |
17 | 17 |
if indent |
18 |
- output.gsub(/\n/i, "\n ").tap { |a| p a } |
|
18 |
+ output.gsub(/\n/i, "\n ") |
|
19 | 19 |
else |
20 | 20 |
output |
21 | 21 |
end |
@@ -12,3 +12,4 @@ EVERNOTE_OAUTH_KEY=evernoteoauthkey |
||
12 | 12 |
EVERNOTE_OAUTH_SECRET=evernoteoauthsecret |
13 | 13 |
FAILED_JOBS_TO_KEEP=2 |
14 | 14 |
REQUIRE_CONFIRMED_EMAIL=false |
15 |
+ENABLE_INSECURE_AGENTS=true |
@@ -0,0 +1,244 @@ |
||
1 |
+require 'rails_helper' |
|
2 |
+ |
|
3 |
+describe Agents::CsvAgent do |
|
4 |
+ before(:each) do |
|
5 |
+ @valid_params = { |
|
6 |
+ 'mode' => 'parse', |
|
7 |
+ 'separator' => ',', |
|
8 |
+ 'use_fields' => '', |
|
9 |
+ 'output' => 'event_per_row', |
|
10 |
+ 'with_header' => 'true', |
|
11 |
+ 'data_path' => '$.data', |
|
12 |
+ 'data_key' => 'data' |
|
13 |
+ } |
|
14 |
+ |
|
15 |
+ @checker = Agents::CsvAgent.new(:name => 'somename', :options => @valid_params) |
|
16 |
+ @checker.user = users(:jane) |
|
17 |
+ @checker.save! |
|
18 |
+ @lfa = Agents::LocalFileAgent.new(name: 'local', options: {path: '{{}}', watch: 'false', append: 'false', mode: 'read'}) |
|
19 |
+ @lfa.user = users(:jane) |
|
20 |
+ @lfa.save! |
|
21 |
+ end |
|
22 |
+ |
|
23 |
+ it_behaves_like 'FileHandlingConsumer' |
|
24 |
+ |
|
25 |
+ context '#validate_options' do |
|
26 |
+ it 'is valid with the given options' do |
|
27 |
+ expect(@checker).to be_valid |
|
28 |
+ end |
|
29 |
+ |
|
30 |
+ it "requires with_header to be either 'true' or 'false'" do |
|
31 |
+ @checker.options['with_header'] = 'true' |
|
32 |
+ expect(@checker).to be_valid |
|
33 |
+ @checker.options['with_header'] = 'false' |
|
34 |
+ expect(@checker).to be_valid |
|
35 |
+ @checker.options['with_header'] = 'test' |
|
36 |
+ expect(@checker).not_to be_valid |
|
37 |
+ end |
|
38 |
+ |
|
39 |
+ it "data_path has to be set in serialize mode" do |
|
40 |
+ @checker.options['mode'] = 'serialize' |
|
41 |
+ @checker.options['data_path'] = '' |
|
42 |
+ expect(@checker).not_to be_valid |
|
43 |
+ end |
|
44 |
+ end |
|
45 |
+ |
|
46 |
+ context '#working' do |
|
47 |
+ it 'is not working without having received an event' do |
|
48 |
+ expect(@checker).not_to be_working |
|
49 |
+ end |
|
50 |
+ |
|
51 |
+ it 'is working after receiving an event without error' do |
|
52 |
+ @checker.last_receive_at = Time.now |
|
53 |
+ expect(@checker).to be_working |
|
54 |
+ end |
|
55 |
+ end |
|
56 |
+ |
|
57 |
+ context '#receive' do |
|
58 |
+ after(:all) do |
|
59 |
+ FileUtils.rm(File.join(Rails.root, 'tmp', 'csv')) |
|
60 |
+ end |
|
61 |
+ |
|
62 |
+ def event_with_contents(contents) |
|
63 |
+ path = File.join(Rails.root, 'tmp', 'csv') |
|
64 |
+ File.open(path, 'w') do |f| |
|
65 |
+ f.write(contents) |
|
66 |
+ end |
|
67 |
+ Event.new(payload: { 'file_pointer' => {'agent_id' => @lfa.id, 'file' => path } }, user_id: @checker.user_id) |
|
68 |
+ end |
|
69 |
+ |
|
70 |
+ context "agent options" do |
|
71 |
+ let(:with_headers) { event_with_contents("one,two\n1,2\n2,3") } |
|
72 |
+ let(:without_headers) { event_with_contents("1,2\n2,3") } |
|
73 |
+ |
|
74 |
+ context "output" do |
|
75 |
+ it "creates one event per row" do |
|
76 |
+ @checker.options['output'] = 'event_per_row' |
|
77 |
+ expect { @checker.receive([with_headers]) }.to change(Event, :count).by(2) |
|
78 |
+ expect(Event.last.payload).to eq(@checker.options['data_key'] => {'one' => '2', 'two' => '3'}) |
|
79 |
+ end |
|
80 |
+ |
|
81 |
+ it "creates one event per file" do |
|
82 |
+ @checker.options['output'] = 'event_per_file' |
|
83 |
+ expect { @checker.receive([with_headers]) }.to change(Event, :count).by(1) |
|
84 |
+ expect(Event.last.payload).to eq(@checker.options['data_key'] => [{"one"=>"1", "two"=>"2"}, {"one"=>"2", "two"=>"3"}]) |
|
85 |
+ end |
|
86 |
+ end |
|
87 |
+ |
|
88 |
+ context "with_header" do |
|
89 |
+ it "works without headers" do |
|
90 |
+ @checker.options['with_header'] = 'false' |
|
91 |
+ expect { @checker.receive([without_headers]) }.to change(Event, :count).by(2) |
|
92 |
+ expect(Event.last.payload).to eq({@checker.options['data_key']=>["2", "3"]}) |
|
93 |
+ end |
|
94 |
+ |
|
95 |
+ it "works without headers and event_per_file" do |
|
96 |
+ @checker.options['with_header'] = 'false' |
|
97 |
+ @checker.options['output'] = 'event_per_file' |
|
98 |
+ expect { @checker.receive([without_headers]) }.to change(Event, :count).by(1) |
|
99 |
+ expect(Event.last.payload).to eq({@checker.options['data_key']=>[['1', '2'], ["2", "3"]]}) |
|
100 |
+ end |
|
101 |
+ end |
|
102 |
+ |
|
103 |
+ context "use_fields" do |
|
104 |
+ it "extracts the specified columns" do |
|
105 |
+ @checker.options['use_fields'] = 'one' |
|
106 |
+ expect { @checker.receive([with_headers]) }.to change(Event, :count).by(2) |
|
107 |
+ expect(Event.last.payload).to eq(@checker.options['data_key'] => {'one' => '2'}) |
|
108 |
+ end |
|
109 |
+ end |
|
110 |
+ |
|
111 |
+ context "data_path" do |
|
112 |
+ it "can receive the CSV via a regular event" do |
|
113 |
+ @checker.options['data_path'] = '$.data' |
|
114 |
+ event = Event.new(payload: {'data' => "one,two\r\n1,2\r\n2,3"}) |
|
115 |
+ expect { @checker.receive([event]) }.to change(Event, :count).by(2) |
|
116 |
+ expect(Event.last.payload).to eq(@checker.options['data_key'] => {'one' => '2', 'two' => '3'}) |
|
117 |
+ end |
|
118 |
+ end |
|
119 |
+ end |
|
120 |
+ |
|
121 |
+ context "handling different CSV formats" do |
|
122 |
+ it "works with windows line endings" do |
|
123 |
+ event = event_with_contents("one,two\r\n1,2\r\n2,3") |
|
124 |
+ expect { @checker.receive([event]) }.to change(Event, :count).by(2) |
|
125 |
+ expect(Event.last.payload).to eq(@checker.options['data_key'] => {'one' => '2', 'two' => '3'}) |
|
126 |
+ end |
|
127 |
+ |
|
128 |
+ it "works with OSX line endings" do |
|
129 |
+ event = event_with_contents("one,two\r1,2\r2,3") |
|
130 |
+ expect { @checker.receive([event]) }.to change(Event, :count).by(2) |
|
131 |
+ expect(Event.last.payload).to eq(@checker.options['data_key'] => {'one' => '2', 'two' => '3'}) |
|
132 |
+ end |
|
133 |
+ |
|
134 |
+ it "handles quotes correctly" do |
|
135 |
+ event = event_with_contents("\"one\",\"two\"\n1,2\n\"\"2, two\",3") |
|
136 |
+ expect { @checker.receive([event]) }.to change(Event, :count).by(2) |
|
137 |
+ expect(Event.last.payload).to eq(@checker.options['data_key'] => {'one' => '"2, two', 'two' => '3'}) |
|
138 |
+ end |
|
139 |
+ |
|
140 |
+ it "works with tab seperated csv" do |
|
141 |
+ event = event_with_contents("one\ttwo\r\n1\t2\r\n2\t3") |
|
142 |
+ @checker.options['separator'] = '\\t' |
|
143 |
+ expect { @checker.receive([event]) }.to change(Event, :count).by(2) |
|
144 |
+ expect(Event.last.payload).to eq(@checker.options['data_key'] => {'one' => '2', 'two' => '3'}) |
|
145 |
+ end |
|
146 |
+ end |
|
147 |
+ |
|
148 |
+ context "serializing" do |
|
149 |
+ before(:each) do |
|
150 |
+ @checker.options['mode'] = 'serialize' |
|
151 |
+ @checker.options['data_path'] = '$.data' |
|
152 |
+ @checker.options['data_key'] = 'data' |
|
153 |
+ end |
|
154 |
+ |
|
155 |
+ it "writes headers when with_header is true" do |
|
156 |
+ event = Event.new(payload: { 'data' => {'key' => 'value', 'key2' => 'value2', 'key3' => 'value3'} }) |
|
157 |
+ expect { @checker.receive([event])}.to change(Event, :count).by(1) |
|
158 |
+ expect(Event.last.payload).to eq('data' => "\"key\",\"key2\",\"key3\"\n\"value\",\"value2\",\"value3\"\n") |
|
159 |
+ end |
|
160 |
+ |
|
161 |
+ it "writes one row per received event" do |
|
162 |
+ event = Event.new(payload: { 'data' => {'key' => 'value', 'key2' => 'value2', 'key3' => 'value3'} }) |
|
163 |
+ event2 = Event.new(payload: { 'data' => {'key' => '2value', 'key2' => '2value2', 'key3' => '2value3'} }) |
|
164 |
+ expect { @checker.receive([event, event2])}.to change(Event, :count).by(1) |
|
165 |
+ expect(Event.last.payload).to eq('data' => "\"key\",\"key2\",\"key3\"\n\"value\",\"value2\",\"value3\"\n\"2value\",\"2value2\",\"2value3\"\n") |
|
166 |
+ end |
|
167 |
+ |
|
168 |
+ it "accepts multiple rows per event" do |
|
169 |
+ event = Event.new(payload: { 'data' => [{'key' => 'value', 'key2' => 'value2', 'key3' => 'value3'}, {'key' => '2value', 'key2' => '2value2', 'key3' => '2value3'}] }) |
|
170 |
+ expect { @checker.receive([event])}.to change(Event, :count).by(1) |
|
171 |
+ expect(Event.last.payload).to eq('data' => "\"key\",\"key2\",\"key3\"\n\"value\",\"value2\",\"value3\"\n\"2value\",\"2value2\",\"2value3\"\n") |
|
172 |
+ end |
|
173 |
+ |
|
174 |
+ it "does not write the headers when with_header is false" do |
|
175 |
+ @checker.options['with_header'] = 'false' |
|
176 |
+ event = Event.new(payload: { 'data' => {'key' => 'value', 'key2' => 'value2', 'key3' => 'value3'} }) |
|
177 |
+ expect { @checker.receive([event])}.to change(Event, :count).by(1) |
|
178 |
+ expect(Event.last.payload).to eq('data' => "\"value\",\"value2\",\"value3\"\n") |
|
179 |
+ end |
|
180 |
+ |
|
181 |
+ it "only serialize the keys specified in use_fields" do |
|
182 |
+ @checker.options['use_fields'] = 'key2, key3' |
|
183 |
+ event = Event.new(payload: { 'data' => {'key' => 'value', 'key2' => 'value2', 'key3' => 'value3'} }) |
|
184 |
+ expect { @checker.receive([event])}.to change(Event, :count).by(1) |
|
185 |
+ expect(Event.last.payload).to eq('data' => "\"key2\",\"key3\"\n\"value2\",\"value3\"\n") |
|
186 |
+ end |
|
187 |
+ |
|
188 |
+ it "respects the order of use_fields" do |
|
189 |
+ @checker.options['use_fields'] = 'key3, key' |
|
190 |
+ event = Event.new(payload: { 'data' => {'key' => 'value', 'key2' => 'value2', 'key3' => 'value3'} }) |
|
191 |
+ expect { @checker.receive([event])}.to change(Event, :count).by(1) |
|
192 |
+ expect(Event.last.payload).to eq('data' => "\"key3\",\"key\"\n\"value3\",\"value\"\n") |
|
193 |
+ end |
|
194 |
+ |
|
195 |
+ it "respects use_fields and writes no header" do |
|
196 |
+ @checker.options['with_header'] = 'false' |
|
197 |
+ @checker.options['use_fields'] = 'key2, key3' |
|
198 |
+ event = Event.new(payload: { 'data' => {'key' => 'value', 'key2' => 'value2', 'key3' => 'value3'} }) |
|
199 |
+ expect { @checker.receive([event])}.to change(Event, :count).by(1) |
|
200 |
+ expect(Event.last.payload).to eq('data' => "\"value2\",\"value3\"\n") |
|
201 |
+ end |
|
202 |
+ |
|
203 |
+ context "arrays" do |
|
204 |
+ it "does not write a header" do |
|
205 |
+ @checker.options['with_header'] = 'false' |
|
206 |
+ event = Event.new(payload: { 'data' => ['value1', 'value2'] }) |
|
207 |
+ event2 = Event.new(payload: { 'data' => ['value3', 'value4'] }) |
|
208 |
+ expect { @checker.receive([event, event2])}.to change(Event, :count).by(1) |
|
209 |
+ expect(Event.last.payload).to eq('data' => "\"value1\",\"value2\"\n\"value3\",\"value4\"\n") |
|
210 |
+ end |
|
211 |
+ |
|
212 |
+ it "handles nested arrays" do |
|
213 |
+ event = Event.new(payload: { 'data' => [['value1', 'value2'], ['value3', 'value4']] }) |
|
214 |
+ expect { @checker.receive([event])}.to change(Event, :count).by(1) |
|
215 |
+ expect(Event.last.payload).to eq('data' => "\"value1\",\"value2\"\n\"value3\",\"value4\"\n") |
|
216 |
+ end |
|
217 |
+ end |
|
218 |
+ end |
|
219 |
+ end |
|
220 |
+ |
|
221 |
+ context '#event_description' do |
|
222 |
+ it "works with event_per_row and headers" do |
|
223 |
+ @checker.options['output'] = 'event_per_row' |
|
224 |
+ @checker.options['with_header'] = 'true' |
|
225 |
+ description = @checker.event_description |
|
226 |
+ expect(description).not_to match(/\n\s+\[\n/) |
|
227 |
+ expect(description).to include(": {\n") |
|
228 |
+ end |
|
229 |
+ |
|
230 |
+ it "works with event_per_file and without headers" do |
|
231 |
+ @checker.options['output'] = 'event_per_file' |
|
232 |
+ @checker.options['with_header'] = 'false' |
|
233 |
+ description = @checker.event_description |
|
234 |
+ expect(description).to match(/\n\s+\[\n/) |
|
235 |
+ expect(description).not_to include(": {\n") |
|
236 |
+ end |
|
237 |
+ |
|
238 |
+ it "shows dummy CSV when in serialize mode" do |
|
239 |
+ @checker.options['mode'] = 'serialize' |
|
240 |
+ description = @checker.event_description |
|
241 |
+ expect(description).to include('"generated\",\"csv') |
|
242 |
+ end |
|
243 |
+ end |
|
244 |
+end |
@@ -8,12 +8,74 @@ describe Agents::FtpsiteAgent do |
||
8 | 8 |
'expected_update_period_in_days' => 1, |
9 | 9 |
'url' => "ftp://ftp.example.org/pub/releases/", |
10 | 10 |
'patterns' => ["example*.tar.gz"], |
11 |
+ 'mode' => 'read', |
|
12 |
+ 'filename' => 'test', |
|
13 |
+ 'data' => '{{ data }}' |
|
11 | 14 |
} |
12 | 15 |
@checker = Agents::FtpsiteAgent.new(:name => "Example", :options => @site, :keep_events_for => 2.days) |
13 | 16 |
@checker.user = users(:bob) |
14 | 17 |
@checker.save! |
15 | 18 |
end |
16 | 19 |
|
20 |
+ context "#validate_options" do |
|
21 |
+ it "requires url to be a valid URI" do |
|
22 |
+ @checker.options['url'] = 'not_valid' |
|
23 |
+ expect(@checker).not_to be_valid |
|
24 |
+ end |
|
25 |
+ |
|
26 |
+ it "allows an URI without a path" do |
|
27 |
+ @checker.options['url'] = 'ftp://ftp.example.org' |
|
28 |
+ expect(@checker).to be_valid |
|
29 |
+ end |
|
30 |
+ |
|
31 |
+ it "does not check the url when liquid output markup is used" do |
|
32 |
+ @checker.options['url'] = 'ftp://{{ ftp_host }}' |
|
33 |
+ expect(@checker).to be_valid |
|
34 |
+ end |
|
35 |
+ |
|
36 |
+ it "requires patterns to be present and not empty array" do |
|
37 |
+ @checker.options['patterns'] = '' |
|
38 |
+ expect(@checker).not_to be_valid |
|
39 |
+ @checker.options['patterns'] = 'not an array' |
|
40 |
+ expect(@checker).not_to be_valid |
|
41 |
+ @checker.options['patterns'] = [] |
|
42 |
+ expect(@checker).not_to be_valid |
|
43 |
+ end |
|
44 |
+ |
|
45 |
+ it "when present timestamp must be parsable into a Time object instance" do |
|
46 |
+ @checker.options['timestamp'] = '2015-01-01 00:00:01' |
|
47 |
+ expect(@checker).to be_valid |
|
48 |
+ @checker.options['timestamp'] = 'error' |
|
49 |
+ expect(@checker).not_to be_valid |
|
50 |
+ end |
|
51 |
+ |
|
52 |
+ it "requires mode to be set to 'read' or 'write'" do |
|
53 |
+ @checker.options['mode'] = 'write' |
|
54 |
+ expect(@checker).to be_valid |
|
55 |
+ @checker.options['mode'] = '' |
|
56 |
+ expect(@checker).not_to be_valid |
|
57 |
+ end |
|
58 |
+ |
|
59 |
+ it 'automatically sets mode to read when the agent is a new record' do |
|
60 |
+ checker = Agents::FtpsiteAgent.new(name: 'test', options: @site.except('mode')) |
|
61 |
+ checker.user = users(:bob) |
|
62 |
+ expect(checker).to be_valid |
|
63 |
+ expect(checker.options['mode']).to eq('read') |
|
64 |
+ end |
|
65 |
+ |
|
66 |
+ it "requires 'filename' in 'write' mode" do |
|
67 |
+ @checker.options['mode'] = 'write' |
|
68 |
+ @checker.options['filename'] = '' |
|
69 |
+ expect(@checker).not_to be_valid |
|
70 |
+ end |
|
71 |
+ |
|
72 |
+ it "requires 'data' in 'write' mode" do |
|
73 |
+ @checker.options['mode'] = 'write' |
|
74 |
+ @checker.options['data'] = '' |
|
75 |
+ expect(@checker).not_to be_valid |
|
76 |
+ end |
|
77 |
+ end |
|
78 |
+ |
|
17 | 79 |
describe "#check" do |
18 | 80 |
|
19 | 81 |
before do |
@@ -42,6 +104,7 @@ describe Agents::FtpsiteAgent do |
||
42 | 104 |
} |
43 | 105 |
|
44 | 106 |
expect(Event.last(2).first.payload).to eq({ |
107 |
+ 'file_pointer' => { 'file' => 'example-1.1.tar.gz', 'agent_id' => @checker.id }, |
|
45 | 108 |
'url' => 'ftp://ftp.example.org/pub/releases/example-1.1.tar.gz', |
46 | 109 |
'filename' => 'example-1.1.tar.gz', |
47 | 110 |
'timestamp' => '2014-04-01T10:00:00Z', |
@@ -71,12 +134,14 @@ describe Agents::FtpsiteAgent do |
||
71 | 134 |
} |
72 | 135 |
|
73 | 136 |
expect(Event.last(2).first.payload).to eq({ |
137 |
+ 'file_pointer' => { 'file' => 'example-1.2.tar.gz', 'agent_id' => @checker.id }, |
|
74 | 138 |
'url' => 'ftp://ftp.example.org/pub/releases/example-1.2.tar.gz', |
75 | 139 |
'filename' => 'example-1.2.tar.gz', |
76 | 140 |
'timestamp' => '2014-04-02T10:00:00Z', |
77 | 141 |
}) |
78 | 142 |
|
79 | 143 |
expect(Event.last.payload).to eq({ |
144 |
+ 'file_pointer' => { 'file' => 'example latest.tar.gz', 'agent_id' => @checker.id }, |
|
80 | 145 |
'url' => 'ftp://ftp.example.org/pub/releases/example%20latest.tar.gz', |
81 | 146 |
'filename' => 'example latest.tar.gz', |
82 | 147 |
'timestamp' => '2014-04-02T10:00:01Z', |
@@ -113,5 +178,83 @@ describe Agents::FtpsiteAgent do |
||
113 | 178 |
end |
114 | 179 |
end |
115 | 180 |
|
181 |
+ context "#open_ftp" do |
|
182 |
+ before(:each) do |
|
183 |
+ @ftp_mock = mock() |
|
184 |
+ mock(@ftp_mock).close |
|
185 |
+ mock(@ftp_mock).connect('ftp.example.org', 21) |
|
186 |
+ mock(@ftp_mock).passive=(true) |
|
187 |
+ mock(Net::FTP).new { @ftp_mock } |
|
188 |
+ end |
|
189 |
+ context 'with_path' do |
|
190 |
+ before(:each) { mock(@ftp_mock).chdir('pub/releases') } |
|
191 |
+ |
|
192 |
+ it "logs in as anonymous when no user and password are given" do |
|
193 |
+ mock(@ftp_mock).login('anonymous', 'anonymous@') |
|
194 |
+ expect { |b| @checker.open_ftp(@checker.base_uri, &b) }.to yield_with_args(@ftp_mock) |
|
195 |
+ end |
|
196 |
+ |
|
197 |
+ it "passes the provided user and password" do |
|
198 |
+ @checker.options['url'] = "ftp://user:password@ftp.example.org/pub/releases/" |
|
199 |
+ mock(@ftp_mock).login('user', 'password') |
|
200 |
+ expect { |b| @checker.open_ftp(@checker.base_uri, &b) }.to yield_with_args(@ftp_mock) |
|
201 |
+ end |
|
202 |
+ end |
|
203 |
+ |
|
204 |
+ it "does not call chdir when no path is given" do |
|
205 |
+ @checker.options['url'] = "ftp://ftp.example.org/" |
|
206 |
+ mock(@ftp_mock).login('anonymous', 'anonymous@') |
|
207 |
+ expect { |b| @checker.open_ftp(@checker.base_uri, &b) }.to yield_with_args(@ftp_mock) |
|
208 |
+ end |
|
209 |
+ end |
|
210 |
+ |
|
211 |
+ context "#get_io" do |
|
212 |
+ it "returns the contents of the file" do |
|
213 |
+ ftp_mock= mock() |
|
214 |
+ mock(ftp_mock).getbinaryfile('file', nil).yields('data') |
|
215 |
+ mock(@checker).open_ftp(@checker.base_uri).yields(ftp_mock) |
|
216 |
+ expect(@checker.get_io('file').read).to eq('data') |
|
217 |
+ end |
|
218 |
+ |
|
219 |
+ it "uses the encoding specified in force_encoding to convert the data to UTF-8" do |
|
220 |
+ ftp_mock= mock() |
|
221 |
+ mock(ftp_mock).getbinaryfile('file', nil).yields('ümlaut'.force_encoding('ISO-8859-15')) |
|
222 |
+ mock(@checker).open_ftp(@checker.base_uri).yields(ftp_mock) |
|
223 |
+ expect(@checker.get_io('file').read).to eq('ümlaut') |
|
224 |
+ end |
|
225 |
+ |
|
226 |
+ it "returns an empty StringIO instance when no data was read" do |
|
227 |
+ ftp_mock= mock() |
|
228 |
+ mock(ftp_mock).getbinaryfile('file', nil) |
|
229 |
+ mock(@checker).open_ftp(@checker.base_uri).yields(ftp_mock) |
|
230 |
+ expect(@checker.get_io('file').length).to eq(0) |
|
231 |
+ end |
|
232 |
+ end |
|
233 |
+ |
|
234 |
+ context "#receive" do |
|
235 |
+ before(:each) do |
|
236 |
+ @checker.options['mode'] = 'write' |
|
237 |
+ @checker.options['filename'] = 'file.txt' |
|
238 |
+ @checker.options['data'] = '{{ data }}' |
|
239 |
+ @ftp_mock= mock() |
|
240 |
+ @stringio = StringIO.new() |
|
241 |
+ mock(@checker).open_ftp(@checker.base_uri).yields(@ftp_mock) |
|
242 |
+ end |
|
243 |
+ |
|
244 |
+ it "writes the data at data into a file" do |
|
245 |
+ mock(StringIO).new('hello world🔥') { @stringio } |
|
246 |
+ mock(@ftp_mock).storbinary('STOR file.txt', @stringio, Net::FTP::DEFAULT_BLOCKSIZE) |
|
247 |
+ event = Event.new(payload: {'data' => 'hello world🔥'}) |
|
248 |
+ @checker.receive([event]) |
|
249 |
+ end |
|
250 |
+ |
|
251 |
+ it "converts the string encoding when force_encoding is specified" do |
|
252 |
+ @checker.options['force_encoding'] = 'ISO-8859-1' |
|
253 |
+ mock(StringIO).new('hello world?') { @stringio } |
|
254 |
+ mock(@ftp_mock).storbinary('STOR file.txt', @stringio, Net::FTP::DEFAULT_BLOCKSIZE) |
|
255 |
+ event = Event.new(payload: {'data' => 'hello world🔥'}) |
|
256 |
+ @checker.receive([event]) |
|
257 |
+ end |
|
258 |
+ end |
|
116 | 259 |
end |
117 | 260 |
end |
@@ -0,0 +1,276 @@ |
||
1 |
+require 'rails_helper' |
|
2 |
+ |
|
3 |
+describe Agents::LocalFileAgent do |
|
4 |
+ before(:each) do |
|
5 |
+ @valid_params = { |
|
6 |
+ 'mode' => 'read', |
|
7 |
+ 'watch' => 'false', |
|
8 |
+ 'append' => 'false', |
|
9 |
+ 'path' => File.join(Rails.root, 'tmp', 'spec') |
|
10 |
+ } |
|
11 |
+ FileUtils.mkdir_p File.join(Rails.root, 'tmp', 'spec') |
|
12 |
+ |
|
13 |
+ @checker = Agents::LocalFileAgent.new(:name => "somename", :options => @valid_params) |
|
14 |
+ @checker.user = users(:jane) |
|
15 |
+ @checker.save! |
|
16 |
+ end |
|
17 |
+ |
|
18 |
+ after(:all) do |
|
19 |
+ FileUtils.rm_r File.join(Rails.root, 'tmp', 'spec') |
|
20 |
+ end |
|
21 |
+ |
|
22 |
+ describe "#validate_options" do |
|
23 |
+ it "is valid with the given options" do |
|
24 |
+ expect(@checker).to be_valid |
|
25 |
+ end |
|
26 |
+ |
|
27 |
+ it "requires mode to be either 'read' or 'write'" do |
|
28 |
+ @checker.options['mode'] = 'write' |
|
29 |
+ expect(@checker).to be_valid |
|
30 |
+ @checker.options['mode'] = 'write' |
|
31 |
+ expect(@checker).to be_valid |
|
32 |
+ @checker.options['mode'] = 'test' |
|
33 |
+ expect(@checker).not_to be_valid |
|
34 |
+ end |
|
35 |
+ |
|
36 |
+ it "requires the path to be set" do |
|
37 |
+ @checker.options['path'] = '' |
|
38 |
+ expect(@checker).not_to be_valid |
|
39 |
+ end |
|
40 |
+ |
|
41 |
+ it "requires watch to be present" do |
|
42 |
+ @checker.options['watch'] = '' |
|
43 |
+ expect(@checker).not_to be_valid |
|
44 |
+ end |
|
45 |
+ |
|
46 |
+ it "requires watch to be either 'true' or 'false'" do |
|
47 |
+ @checker.options['watch'] = 'true' |
|
48 |
+ expect(@checker).to be_valid |
|
49 |
+ @checker.options['watch'] = 'false' |
|
50 |
+ expect(@checker).to be_valid |
|
51 |
+ @checker.options['watch'] = 'test' |
|
52 |
+ expect(@checker).not_to be_valid |
|
53 |
+ end |
|
54 |
+ |
|
55 |
+ it "requires append to be either 'true' or 'false'" do |
|
56 |
+ @checker.options['append'] = 'true' |
|
57 |
+ expect(@checker).to be_valid |
|
58 |
+ @checker.options['append'] = 'false' |
|
59 |
+ expect(@checker).to be_valid |
|
60 |
+ @checker.options['append'] = 'test' |
|
61 |
+ expect(@checker).not_to be_valid |
|
62 |
+ end |
|
63 |
+ end |
|
64 |
+ |
|
65 |
+ context "#working" do |
|
66 |
+ it "is working with no recent errors in read mode" do |
|
67 |
+ @checker.last_check_at = Time.now |
|
68 |
+ expect(@checker).to be_working |
|
69 |
+ end |
|
70 |
+ |
|
71 |
+ it "is working with no recent errors in write mode" do |
|
72 |
+ @checker.options['mode'] = 'write' |
|
73 |
+ @checker.last_receive_at = Time.now |
|
74 |
+ expect(@checker).to be_working |
|
75 |
+ end |
|
76 |
+ end |
|
77 |
+ |
|
78 |
+ context "#check_path_existance" do |
|
79 |
+ it "is truethy when the path exists" do |
|
80 |
+ expect(@checker.check_path_existance).to be_truthy |
|
81 |
+ end |
|
82 |
+ |
|
83 |
+ it "is falsy when the path does not exist" do |
|
84 |
+ @checker.options['path'] = '/doesnotexist' |
|
85 |
+ expect(@checker.check_path_existance).to be_falsy |
|
86 |
+ end |
|
87 |
+ |
|
88 |
+ it "create a log entry" do |
|
89 |
+ @checker.options['path'] = '/doesnotexist' |
|
90 |
+ expect { @checker.check_path_existance(true) }.to change(AgentLog, :count).by(1) |
|
91 |
+ end |
|
92 |
+ |
|
93 |
+ it "works with non-expanded paths" do |
|
94 |
+ @checker.options['path'] = '~' |
|
95 |
+ expect(@checker.check_path_existance).to be_truthy |
|
96 |
+ end |
|
97 |
+ end |
|
98 |
+ |
|
99 |
+ def with_files(*files) |
|
100 |
+ files.each { |f| FileUtils.touch(f) } |
|
101 |
+ yield |
|
102 |
+ files.each { |f| FileUtils.rm(f) } |
|
103 |
+ end |
|
104 |
+ |
|
105 |
+ context "#check" do |
|
106 |
+ it "does not create events when the directory is empty" do |
|
107 |
+ expect { @checker.check }.to change(Event, :count).by(0) |
|
108 |
+ end |
|
109 |
+ |
|
110 |
+ it "creates an event for every file in the directory" do |
|
111 |
+ with_files(File.join(Rails.root, 'tmp', 'spec', 'one'), File.join(Rails.root, 'tmp', 'spec', 'two')) do |
|
112 |
+ expect { @checker.check }.to change(Event, :count).by(2) |
|
113 |
+ expect(Event.last.payload.has_key?('file_pointer')).to be_truthy |
|
114 |
+ end |
|
115 |
+ end |
|
116 |
+ |
|
117 |
+ it "creates an event if the configured file exists" do |
|
118 |
+ @checker.options['path'] = File.join(Rails.root, 'tmp', 'spec', 'one') |
|
119 |
+ with_files(File.join(Rails.root, 'tmp', 'spec', 'one'), File.join(Rails.root, 'tmp', 'spec', 'two')) do |
|
120 |
+ expect { @checker.check }.to change(Event, :count).by(1) |
|
121 |
+ payload = Event.last.payload |
|
122 |
+ expect(payload.has_key?('file_pointer')).to be_truthy |
|
123 |
+ expect(payload['file_pointer']['file']).to eq(@checker.options['path']) |
|
124 |
+ end |
|
125 |
+ end |
|
126 |
+ |
|
127 |
+ it "does not run when ENABLE_INSECURE_AGENTS is not set to true" do |
|
128 |
+ ENV['ENABLE_INSECURE_AGENTS'] = 'false' |
|
129 |
+ expect { @checker.check }.to change(AgentLog, :count).by(1) |
|
130 |
+ ENV['ENABLE_INSECURE_AGENTS'] = 'true' |
|
131 |
+ end |
|
132 |
+ end |
|
133 |
+ |
|
134 |
+ context "#event_description" do |
|
135 |
+ it "should include event_type when watch is set to true" do |
|
136 |
+ @checker.options['watch'] = 'true' |
|
137 |
+ expect(@checker.event_description).to include('event_type') |
|
138 |
+ end |
|
139 |
+ |
|
140 |
+ it "should not include event_type when watch is set to false" do |
|
141 |
+ @checker.options['watch'] = 'false' |
|
142 |
+ expect(@checker.event_description).not_to include('event_type') |
|
143 |
+ end |
|
144 |
+ end |
|
145 |
+ |
|
146 |
+ it "get_io opens the file" do |
|
147 |
+ mock(File).open('test', 'r') |
|
148 |
+ @checker.get_io('test') |
|
149 |
+ end |
|
150 |
+ |
|
151 |
+ context "#start_worker?" do |
|
152 |
+ it "reeturns true when watch is true" do |
|
153 |
+ @checker.options['watch'] = 'true' |
|
154 |
+ expect(@checker.start_worker?).to be_truthy |
|
155 |
+ end |
|
156 |
+ |
|
157 |
+ it "returns false when watch is false" do |
|
158 |
+ @checker.options['watch'] = 'false' |
|
159 |
+ expect(@checker.start_worker?).to be_falsy |
|
160 |
+ end |
|
161 |
+ end |
|
162 |
+ |
|
163 |
+ context "#receive" do |
|
164 |
+ before(:each) do |
|
165 |
+ @checker.options['mode'] = 'write' |
|
166 |
+ @checker.options['data'] = '{{ data }}' |
|
167 |
+ @file_mock = mock() |
|
168 |
+ end |
|
169 |
+ |
|
170 |
+ it "writes the data at data into a file" do |
|
171 |
+ mock(@file_mock).write('hello world') |
|
172 |
+ event = Event.new(payload: {'data' => 'hello world'}) |
|
173 |
+ mock(File).open(File.join(Rails.root, 'tmp', 'spec'), 'w').yields @file_mock |
|
174 |
+ @checker.receive([event]) |
|
175 |
+ end |
|
176 |
+ |
|
177 |
+ it "appends the data at data onto a file" do |
|
178 |
+ mock(@file_mock).write('hello world') |
|
179 |
+ @checker.options['append'] = 'true' |
|
180 |
+ event = Event.new(payload: {'data' => 'hello world'}) |
|
181 |
+ mock(File).open(File.join(Rails.root, 'tmp', 'spec'), 'a').yields @file_mock |
|
182 |
+ @checker.receive([event]) |
|
183 |
+ end |
|
184 |
+ |
|
185 |
+ it "does not receive when ENABLE_INSECURE_AGENTS is not set to true" do |
|
186 |
+ ENV['ENABLE_INSECURE_AGENTS'] = 'false' |
|
187 |
+ expect { @checker.receive([]) }.to change(AgentLog, :count).by(1) |
|
188 |
+ ENV['ENABLE_INSECURE_AGENTS'] = 'true' |
|
189 |
+ end |
|
190 |
+ end |
|
191 |
+ |
|
192 |
+ describe describe Agents::LocalFileAgent::Worker do |
|
193 |
+ require 'listen' |
|
194 |
+ |
|
195 |
+ before(:each) do |
|
196 |
+ @checker.options['watch'] = true |
|
197 |
+ @checker.save |
|
198 |
+ @worker = Agents::LocalFileAgent::Worker.new(agent: @checker) |
|
199 |
+ @listen_mock = mock() |
|
200 |
+ end |
|
201 |
+ |
|
202 |
+ context "#setup" do |
|
203 |
+ it "initializes the listen gem" do |
|
204 |
+ mock(Listen).to(@checker.options['path'], ignore!: []) |
|
205 |
+ @worker.setup |
|
206 |
+ end |
|
207 |
+ end |
|
208 |
+ |
|
209 |
+ context "#run" do |
|
210 |
+ before(:each) do |
|
211 |
+ stub(Listen).to { @listen_mock } |
|
212 |
+ @worker.setup |
|
213 |
+ end |
|
214 |
+ |
|
215 |
+ it "starts to listen to changes in the directory when the path is present" do |
|
216 |
+ mock(@worker).sleep |
|
217 |
+ mock(@listen_mock).start |
|
218 |
+ @worker.run |
|
219 |
+ end |
|
220 |
+ |
|
221 |
+ it "does nothing when the path does not exist" do |
|
222 |
+ mock(@worker.agent).check_path_existance(true) { false } |
|
223 |
+ dont_allow(@listen_mock).start |
|
224 |
+ mock(@worker).sleep { raise "Sleeping" } |
|
225 |
+ expect { @worker.run }.to raise_exception(RuntimeError, 'Sleeping') |
|
226 |
+ end |
|
227 |
+ end |
|
228 |
+ |
|
229 |
+ context "#stop" do |
|
230 |
+ it "stops the listen gem" do |
|
231 |
+ stub(Listen).to { @listen_mock } |
|
232 |
+ @worker.setup |
|
233 |
+ mock(@listen_mock).stop |
|
234 |
+ @worker.stop |
|
235 |
+ end |
|
236 |
+ end |
|
237 |
+ |
|
238 |
+ context "#callback" do |
|
239 |
+ let(:file) { File.join(Rails.root, 'tmp', 'one') } |
|
240 |
+ let(:file2) { File.join(Rails.root, 'tmp', 'one2') } |
|
241 |
+ |
|
242 |
+ it "creates an event for modifies files" do |
|
243 |
+ expect { @worker.send(:callback, [file], [], [])}.to change(Event, :count).by(1) |
|
244 |
+ payload = Event.last.payload |
|
245 |
+ expect(payload['event_type']).to eq('modified') |
|
246 |
+ end |
|
247 |
+ |
|
248 |
+ it "creates an event for modifies files" do |
|
249 |
+ expect { @worker.send(:callback, [], [file], [])}.to change(Event, :count).by(1) |
|
250 |
+ payload = Event.last.payload |
|
251 |
+ expect(payload['event_type']).to eq('added') |
|
252 |
+ end |
|
253 |
+ |
|
254 |
+ it "creates an event for modifies files" do |
|
255 |
+ expect { @worker.send(:callback, [], [], [file])}.to change(Event, :count).by(1) |
|
256 |
+ payload = Event.last.payload |
|
257 |
+ expect(payload['event_type']).to eq('removed') |
|
258 |
+ end |
|
259 |
+ |
|
260 |
+ it "creates an event each changed file" do |
|
261 |
+ expect { @worker.send(:callback, [], [file], [file2])}.to change(Event, :count).by(2) |
|
262 |
+ end |
|
263 |
+ end |
|
264 |
+ |
|
265 |
+ context "#listen_options" do |
|
266 |
+ it "returns the path when a directory is given" do |
|
267 |
+ expect(@worker.send(:listen_options)).to eq([File.join(Rails.root, 'tmp', 'spec'), ignore!: []]) |
|
268 |
+ end |
|
269 |
+ |
|
270 |
+ it "restricts to only the specified filename" do |
|
271 |
+ @worker.agent.options['path'] = File.join(Rails.root, 'tmp', 'one') |
|
272 |
+ expect(@worker.send(:listen_options)).to eq([File.join(Rails.root, 'tmp'), { only: /\Aone\z/, ignore!: [] } ]) |
|
273 |
+ end |
|
274 |
+ end |
|
275 |
+ end |
|
276 |
+end |
@@ -0,0 +1,47 @@ |
||
1 |
+require 'rails_helper' |
|
2 |
+ |
|
3 |
+describe Agents::ReadFileAgent do |
|
4 |
+ before(:each) do |
|
5 |
+ @valid_params = { |
|
6 |
+ 'data_key' => 'data', |
|
7 |
+ } |
|
8 |
+ |
|
9 |
+ @checker = Agents::ReadFileAgent.new(:name => 'somename', :options => @valid_params) |
|
10 |
+ @checker.user = users(:jane) |
|
11 |
+ @checker.save! |
|
12 |
+ end |
|
13 |
+ |
|
14 |
+ it_behaves_like 'FileHandlingConsumer' |
|
15 |
+ |
|
16 |
+ context '#validate_options' do |
|
17 |
+ it 'is valid with the given options' do |
|
18 |
+ expect(@checker).to be_valid |
|
19 |
+ end |
|
20 |
+ |
|
21 |
+ it "requires data_key to be present" do |
|
22 |
+ @checker.options['data_key'] = '' |
|
23 |
+ expect(@checker).not_to be_valid |
|
24 |
+ end |
|
25 |
+ end |
|
26 |
+ |
|
27 |
+ context '#working' do |
|
28 |
+ it 'is not working without having received an event' do |
|
29 |
+ expect(@checker).not_to be_working |
|
30 |
+ end |
|
31 |
+ |
|
32 |
+ it 'is working after receiving an event without error' do |
|
33 |
+ @checker.last_receive_at = Time.now |
|
34 |
+ expect(@checker).to be_working |
|
35 |
+ end |
|
36 |
+ end |
|
37 |
+ |
|
38 |
+ context '#receive' do |
|
39 |
+ it "emits an event with the contents of the receives files" do |
|
40 |
+ event = Event.new(payload: {file_pointer: {agent_id: 111, file: 'test'}}) |
|
41 |
+ io_mock = mock() |
|
42 |
+ mock(@checker).get_io(event) { StringIO.new("testdata") } |
|
43 |
+ expect { @checker.receive([event]) }.to change(Event, :count).by(1) |
|
44 |
+ expect(Event.last.payload).to eq('data' => 'testdata') |
|
45 |
+ end |
|
46 |
+ end |
|
47 |
+end |
@@ -0,0 +1,220 @@ |
||
1 |
+require 'rails_helper' |
|
2 |
+ |
|
3 |
+describe Agents::S3Agent do |
|
4 |
+ before(:each) do |
|
5 |
+ @valid_params = { |
|
6 |
+ 'mode' => 'read', |
|
7 |
+ 'access_key_id' => '32343242', |
|
8 |
+ 'access_key_secret' => '1231312', |
|
9 |
+ 'watch' => 'false', |
|
10 |
+ 'bucket' => 'testbucket', |
|
11 |
+ 'region' => 'us-east-1', |
|
12 |
+ 'filename' => 'test.txt', |
|
13 |
+ 'data' => '{{ data }}' |
|
14 |
+ } |
|
15 |
+ |
|
16 |
+ @checker = Agents::S3Agent.new(:name => "somename", :options => @valid_params) |
|
17 |
+ @checker.user = users(:jane) |
|
18 |
+ @checker.save! |
|
19 |
+ end |
|
20 |
+ |
|
21 |
+ describe "#validate_options" do |
|
22 |
+ it "requires the bucket to be set" do |
|
23 |
+ @checker.options['bucket'] = '' |
|
24 |
+ expect(@checker).not_to be_valid |
|
25 |
+ end |
|
26 |
+ |
|
27 |
+ it "requires watch to be present" do |
|
28 |
+ @checker.options['watch'] = '' |
|
29 |
+ expect(@checker).not_to be_valid |
|
30 |
+ end |
|
31 |
+ |
|
32 |
+ it "requires watch to be either 'true' or 'false'" do |
|
33 |
+ @checker.options['watch'] = 'true' |
|
34 |
+ expect(@checker).to be_valid |
|
35 |
+ @checker.options['watch'] = 'false' |
|
36 |
+ expect(@checker).to be_valid |
|
37 |
+ @checker.options['watch'] = 'test' |
|
38 |
+ expect(@checker).not_to be_valid |
|
39 |
+ end |
|
40 |
+ |
|
41 |
+ it "requires region to be present" do |
|
42 |
+ @checker.options['region'] = '' |
|
43 |
+ expect(@checker).not_to be_valid |
|
44 |
+ end |
|
45 |
+ |
|
46 |
+ it "requires mode to be set to 'read' or 'write'" do |
|
47 |
+ @checker.options['mode'] = 'write' |
|
48 |
+ expect(@checker).to be_valid |
|
49 |
+ @checker.options['mode'] = '' |
|
50 |
+ expect(@checker).not_to be_valid |
|
51 |
+ end |
|
52 |
+ |
|
53 |
+ it "requires 'filename' in 'write' mode" do |
|
54 |
+ @checker.options['mode'] = 'write' |
|
55 |
+ @checker.options['filename'] = '' |
|
56 |
+ expect(@checker).not_to be_valid |
|
57 |
+ end |
|
58 |
+ |
|
59 |
+ it "requires 'data' in 'write' mode" do |
|
60 |
+ @checker.options['mode'] = 'write' |
|
61 |
+ @checker.options['data'] = '' |
|
62 |
+ expect(@checker).not_to be_valid |
|
63 |
+ end |
|
64 |
+ end |
|
65 |
+ |
|
66 |
+ describe "#validating" do |
|
67 |
+ it "validates the key" do |
|
68 |
+ mock(@checker).client { raise Aws::S3::Errors::SignatureDoesNotMatch.new('', '') } |
|
69 |
+ expect(@checker.validate_access_key_id).to be_falsy |
|
70 |
+ end |
|
71 |
+ |
|
72 |
+ it "validates the secret" do |
|
73 |
+ mock(@checker).buckets { true } |
|
74 |
+ expect(@checker.validate_access_key_secret).to be_truthy |
|
75 |
+ end |
|
76 |
+ end |
|
77 |
+ |
|
78 |
+ it "completes the buckets" do |
|
79 |
+ mock(@checker).buckets { [OpenStruct.new(name: 'test'), OpenStruct.new(name: 'test2')]} |
|
80 |
+ expect(@checker.complete_bucket).to eq([{text: 'test', id: 'test'}, {text: 'test2', id: 'test2'}]) |
|
81 |
+ end |
|
82 |
+ |
|
83 |
+ context "#working" do |
|
84 |
+ it "is working with no recent errors" do |
|
85 |
+ @checker.last_check_at = Time.now |
|
86 |
+ expect(@checker).to be_working |
|
87 |
+ end |
|
88 |
+ end |
|
89 |
+ |
|
90 |
+ context "#check" do |
|
91 |
+ context "not watching" do |
|
92 |
+ it "emits an event for every file" do |
|
93 |
+ mock(@checker).get_bucket_contents { {"test"=>"231232", "test2"=>"4564545"} } |
|
94 |
+ expect { @checker.check }.to change(Event, :count).by(2) |
|
95 |
+ expect(Event.last.payload).to eq({"file_pointer" => {"file"=>"test2", "agent_id"=> @checker.id}}) |
|
96 |
+ end |
|
97 |
+ end |
|
98 |
+ |
|
99 |
+ context "watching" do |
|
100 |
+ before(:each) do |
|
101 |
+ @checker.options['watch'] = 'true' |
|
102 |
+ end |
|
103 |
+ |
|
104 |
+ it "does not emit any events on the first run" do |
|
105 |
+ contents = {"test"=>"231232", "test2"=>"4564545"} |
|
106 |
+ mock(@checker).get_bucket_contents { contents } |
|
107 |
+ expect { @checker.check }.not_to change(Event, :count) |
|
108 |
+ expect(@checker.memory).to eq('seen_contents' => contents) |
|
109 |
+ end |
|
110 |
+ |
|
111 |
+ context "detecting changes" do |
|
112 |
+ before(:each) do |
|
113 |
+ contents = {"test"=>"231232", "test2"=>"4564545"} |
|
114 |
+ mock(@checker).get_bucket_contents { contents } |
|
115 |
+ expect { @checker.check }.not_to change(Event, :count) |
|
116 |
+ @checker.last_check_at = Time.now |
|
117 |
+ end |
|
118 |
+ |
|
119 |
+ it "emits events for removed files" do |
|
120 |
+ contents = {"test"=>"231232"} |
|
121 |
+ mock(@checker).get_bucket_contents { contents } |
|
122 |
+ expect { @checker.check }.to change(Event, :count).by(1) |
|
123 |
+ expect(Event.last.payload).to eq({"file_pointer" => {"file" => "test2", "agent_id"=> @checker.id}, "event_type" => "removed"}) |
|
124 |
+ end |
|
125 |
+ |
|
126 |
+ it "emits events for modified files" do |
|
127 |
+ contents = {"test"=>"231232", "test2"=>"changed"} |
|
128 |
+ mock(@checker).get_bucket_contents { contents } |
|
129 |
+ expect { @checker.check }.to change(Event, :count).by(1) |
|
130 |
+ expect(Event.last.payload).to eq({"file_pointer" => {"file" => "test2", "agent_id"=> @checker.id}, "event_type" => "modified"}) |
|
131 |
+ end |
|
132 |
+ it "emits events for added files" do |
|
133 |
+ contents = {"test"=>"231232", "test2"=>"4564545", "test3" => "31231231"} |
|
134 |
+ mock(@checker).get_bucket_contents { contents } |
|
135 |
+ expect { @checker.check }.to change(Event, :count).by(1) |
|
136 |
+ expect(Event.last.payload).to eq({"file_pointer" => {"file" => "test3", "agent_id"=> @checker.id}, "event_type" => "added"}) |
|
137 |
+ end |
|
138 |
+ end |
|
139 |
+ |
|
140 |
+ context "error handling" do |
|
141 |
+ it "handles AccessDenied exceptions" do |
|
142 |
+ mock(@checker).get_bucket_contents { raise Aws::S3::Errors::AccessDenied.new('', '') } |
|
143 |
+ expect { @checker.check }.to change(AgentLog, :count).by(1) |
|
144 |
+ expect(AgentLog.last.message).to eq("Could not access 'testbucket' Aws::S3::Errors::AccessDenied ") |
|
145 |
+ end |
|
146 |
+ |
|
147 |
+ it "handles generic S3 exceptions" do |
|
148 |
+ mock(@checker).get_bucket_contents { raise Aws::S3::Errors::PermanentRedirect.new('', 'error') } |
|
149 |
+ expect { @checker.check }.to change(AgentLog, :count).by(1) |
|
150 |
+ expect(AgentLog.last.message).to eq("Aws::S3::Errors::PermanentRedirect: error") |
|
151 |
+ end |
|
152 |
+ end |
|
153 |
+ end |
|
154 |
+ end |
|
155 |
+ |
|
156 |
+ it "get_io returns a StringIO object" do |
|
157 |
+ stringio =StringIO.new |
|
158 |
+ mock_response = mock() |
|
159 |
+ mock(mock_response).body { stringio } |
|
160 |
+ mock_client = mock() |
|
161 |
+ mock(mock_client).get_object(bucket: 'testbucket', key: 'testfile') { mock_response } |
|
162 |
+ mock(@checker).client { mock_client } |
|
163 |
+ @checker.get_io('testfile') |
|
164 |
+ end |
|
165 |
+ |
|
166 |
+ context "#get_bucket_contents" do |
|
167 |
+ it "returns a hash with the contents of the bucket" do |
|
168 |
+ mock_response = mock() |
|
169 |
+ mock(mock_response).contents { [OpenStruct.new(key: 'test', etag: '231232'), OpenStruct.new(key: 'test2', etag: '4564545')] } |
|
170 |
+ mock_client = mock() |
|
171 |
+ mock(mock_client).list_objects(bucket: 'testbucket') { [mock_response] } |
|
172 |
+ mock(@checker).client { mock_client } |
|
173 |
+ expect(@checker.send(:get_bucket_contents)).to eq({"test"=>"231232", "test2"=>"4564545"}) |
|
174 |
+ end |
|
175 |
+ end |
|
176 |
+ |
|
177 |
+ context "#client" do |
|
178 |
+ it "initializes the S3 client correctly" do |
|
179 |
+ mock_credential = mock() |
|
180 |
+ mock(Aws::Credentials).new('32343242', '1231312') { mock_credential } |
|
181 |
+ mock(Aws::S3::Client).new(credentials: mock_credential, |
|
182 |
+ region: 'us-east-1') |
|
183 |
+ @checker.send(:client) |
|
184 |
+ end |
|
185 |
+ end |
|
186 |
+ |
|
187 |
+ context "#event_description" do |
|
188 |
+ it "should include event_type when watch is set to true" do |
|
189 |
+ @checker.options['watch'] = 'true' |
|
190 |
+ expect(@checker.event_description).to include('event_type') |
|
191 |
+ end |
|
192 |
+ |
|
193 |
+ it "should not include event_type when watch is set to false" do |
|
194 |
+ @checker.options['watch'] = 'false' |
|
195 |
+ expect(@checker.event_description).not_to include('event_type') |
|
196 |
+ end |
|
197 |
+ end |
|
198 |
+ |
|
199 |
+ context "#receive" do |
|
200 |
+ before(:each) do |
|
201 |
+ @checker.options['mode'] = 'write' |
|
202 |
+ @checker.options['filename'] = 'file.txt' |
|
203 |
+ @checker.options['data'] = '{{ data }}' |
|
204 |
+ end |
|
205 |
+ |
|
206 |
+ it "writes the data at data into a file" do |
|
207 |
+ client_mock = mock() |
|
208 |
+ mock(client_mock).put_object(bucket: @checker.options['bucket'], key: @checker.options['filename'], body: 'hello world!') |
|
209 |
+ mock(@checker).client { client_mock } |
|
210 |
+ event = Event.new(payload: {'data' => 'hello world!'}) |
|
211 |
+ @checker.receive([event]) |
|
212 |
+ end |
|
213 |
+ |
|
214 |
+ it "does nothing when mode is set to 'read'" do |
|
215 |
+ @checker.options['mode'] = 'read' |
|
216 |
+ event = Event.new(payload: {'data' => 'hello world!'}) |
|
217 |
+ @checker.receive([event]) |
|
218 |
+ end |
|
219 |
+ end |
|
220 |
+end |
@@ -0,0 +1,16 @@ |
||
1 |
+require 'rails_helper' |
|
2 |
+ |
|
3 |
+shared_examples_for 'FileHandlingConsumer' do |
|
4 |
+ it 'returns a file pointer' do |
|
5 |
+ expect(@checker.get_file_pointer('testfile')).to eq(file_pointer: { file: "testfile", agent_id: @checker.id}) |
|
6 |
+ end |
|
7 |
+ |
|
8 |
+ it 'get_io raises an exception when trying to access an agent of a different user' do |
|
9 |
+ @checker2 = @checker.dup |
|
10 |
+ @checker2.user = users(:bob) |
|
11 |
+ @checker2.save! |
|
12 |
+ expect(@checker2.user.id).not_to eq(@checker.user.id) |
|
13 |
+ event = Event.new(user: @checker.user, payload: {'file_pointer' => {'file' => 'test', 'agent_id' => @checker2.id}}) |
|
14 |
+ expect { @checker.get_io(event) }.to raise_error(ActiveRecord::RecordNotFound) |
|
15 |
+ end |
|
16 |
+end |
@@ -50,4 +50,28 @@ shared_examples_for WorkingHelpers do |
||
50 | 50 |
expect(@agent.received_event_without_error?).to eq(true) |
51 | 51 |
end |
52 | 52 |
end |
53 |
+ |
|
54 |
+ describe "checked_without_error?" do |
|
55 |
+ before do |
|
56 |
+ @agent = described_class.new |
|
57 |
+ end |
|
58 |
+ |
|
59 |
+ it "should return false until the first time check ran" do |
|
60 |
+ expect(@agent.checked_without_error?).to eq(false) |
|
61 |
+ @agent.last_check_at = Time.now |
|
62 |
+ expect(@agent.checked_without_error?).to eq(true) |
|
63 |
+ end |
|
64 |
+ |
|
65 |
+ it "should return false when the last error occured after the check" do |
|
66 |
+ @agent.last_check_at = Time.now - 1.minute |
|
67 |
+ @agent.last_error_log_at = Time.now |
|
68 |
+ expect(@agent.checked_without_error?).to eq(false) |
|
69 |
+ end |
|
70 |
+ |
|
71 |
+ it "should return true when the last check occured after the last error" do |
|
72 |
+ @agent.last_check_at = Time.now |
|
73 |
+ @agent.last_error_log_at = Time.now - 1.minute |
|
74 |
+ expect(@agent.checked_without_error?).to eq(true) |
|
75 |
+ end |
|
76 |
+ end |
|
53 | 77 |
end |