|
1 | 1 | #!/usr/bin/env python
|
2 | 2 | """
|
3 | 3 | title: A CLI tool for exporting data from Elasticsearch into a CSV file.
|
4 |
| -description: Command line utility, written in Python, for querying Elasticsearch in Lucene query syntax and for exporting documents into a CSV file. |
| 4 | +description: Command line utility, written in Python, for querying Elasticsearch in Lucene query syntax or Query DSL syntax and exporting result as documents into a CSV file. |
5 | 5 | usage: es2csv -q '*' -i _all -o ~/file.csv -k -m 100
|
| 6 | + es2csv -q '{"query": {"match_all": {}}}' -r -i _all -o ~/file.csv |
6 | 7 | es2csv -q '*' -i logstash-2015-01-* -f host status message -o ~/file.csv
|
7 | 8 | es2csv -q 'host: localhost' -i logstash-2015-01-01 logstash-2015-01-02 -f host status message -o ~/file.csv
|
8 | 9 | es2csv -q 'host: localhost AND status: GET' -u http://kibana.com:80/es/ -o ~/file.csv
|
9 | 10 | es2csv -q '*' -t dev prod -u http://login:[email protected]:6666/es/ -o ~/file.csv
|
| 11 | + es2csv -q '{"query": {"match_all": {}}, "filter":{"term": {"tags": "dev"}}}' -r -u http://login:[email protected]:6666/es/ -o ~/file.csv |
10 | 12 | """
|
11 | 13 | import os
|
12 | 14 | import sys
|
@@ -85,25 +87,28 @@ def check_indexes(self):
|
85 | 87 |
|
86 | 88 | @retry(elasticsearch.exceptions.ConnectionError, tries=TIMES_TO_TRY)
|
87 | 89 | def search_query(self):
|
88 |
| - query = self.opts.query if not self.opts.tags else '%s AND tags:%s' % ( |
89 |
| - self.opts.query, '(%s)' % ' AND '.join(self.opts.tags)) |
90 |
| - |
91 |
| - if self.opts.debug_mode: |
92 |
| - print('Using these indices: %s' % ', '.join(self.opts.index_prefixes)) |
93 |
| - print('Query: %s' % query) |
94 |
| - print('Output field(s): %s' % ', '.join(self.opts.fields)) |
95 |
| - |
96 | 90 | search_args = dict(
|
97 | 91 | index=','.join(self.opts.index_prefixes),
|
98 |
| - q=query, |
99 | 92 | search_type='scan',
|
100 | 93 | scroll=self.scroll_time,
|
101 | 94 | size=self.scroll_size
|
102 | 95 | )
|
| 96 | + if self.opts.raw_query: |
| 97 | + query = json.loads(self.opts.query) |
| 98 | + search_args['body'] = query |
| 99 | + else: |
| 100 | + query = self.opts.query if not self.opts.tags else '%s AND tags:%s' % ( |
| 101 | + self.opts.query, '(%s)' % ' AND '.join(self.opts.tags)) |
| 102 | + search_args['q'] = query |
103 | 103 |
|
104 | 104 | if '_all' not in self.opts.fields:
|
105 | 105 | search_args['fields'] = ','.join(self.opts.fields)
|
106 | 106 |
|
| 107 | + if self.opts.debug_mode: |
| 108 | + print('Using these indices: %s' % ', '.join(self.opts.index_prefixes)) |
| 109 | + print('Query[%s]: %s' % (('Query DSL', json.dumps(query)) if self.opts.raw_query else ('Lucene', query))) |
| 110 | + print('Output field(s): %s' % ', '.join(self.opts.fields)) |
| 111 | + |
107 | 112 | res = self.es_conn.search(**search_args)
|
108 | 113 |
|
109 | 114 | self.scroll_ids.append(res['_scroll_id'])
|
@@ -236,6 +241,7 @@ def main():
|
236 | 241 | p.add_argument('-d', '--delimiter', dest='delimiter', default=',', type=str, help='Delimiter to use in CSV file. Default is "%(default)s".')
|
237 | 242 | p.add_argument('-m', '--max', dest='max_results', default=0, type=int, metavar='INTEGER', help='Maximum number of results to return. Default is %(default)s.')
|
238 | 243 | p.add_argument('-k', '--kibana_nested', dest='kibana_nested', action='store_true', help='Format nested fields in Kibana style.')
|
| 244 | + p.add_argument('-r', '--raw_query', dest='raw_query', action='store_true', help='Switch query format in the Query DSL.') |
239 | 245 | p.add_argument('-v', '--version', action='version', version='%(prog)s ' + __version__, help='Show version and exit.')
|
240 | 246 | p.add_argument('--debug', dest='debug_mode', action='store_true', help='Debug mode on.')
|
241 | 247 |
|
|
0 commit comments