source: other-projects/hathitrust/wcsa/extracted-features-solr/trunk/solr-ingest/src/main/java/org/hathitrust/extractedfeatures/PerVolumeJSON.java@ 31505

Last change on this file since 31505 was 31505, checked in by davidb, 7 years ago

Added in storing of top-level document metadata as separate solr-doc

  • Property svn:executable set to *
File size: 9.1 KB
Line 
1package org.hathitrust.extractedfeatures;
2
3import java.io.IOException;
4import java.util.ArrayList;
5import java.util.HashMap;
6import java.util.Iterator;
7
8import org.apache.hadoop.io.Text;
9import org.apache.spark.api.java.function.FlatMapFunction;
10import org.apache.spark.api.java.function.Function;
11import org.apache.spark.api.java.function.VoidFunction;
12import org.apache.spark.util.DoubleAccumulator;
13import org.json.JSONArray;
14import org.json.JSONObject;
15
16/*
17class PagedJSON implements Function<String, Boolean> {
18
19 private static final long serialVersionUID = 1L;
20
21 public Boolean call(String s) { return s.contains("a"); }
22}
23 */
24
25
26//public class PerVolumeJSON implements VoidFunction<String>
27public class PerVolumeJSON implements Function<Text,Integer>
28{
29 private static final long serialVersionUID = 1L;
30 protected String _input_dir;
31 protected String _whitelist_filename;
32 protected String _langmap_directory;
33
34 protected final ArrayList<String> _solr_endpoints;
35 protected final int _solr_endpoints_len;
36
37 //protected String _solr_url;
38 protected String _output_dir;
39
40 protected int _verbosity;
41
42 protected WhitelistBloomFilter _whitelist_bloomfilter;
43 protected UniversalPOSLangMap _universal_langmap;
44
45 boolean _icu_tokenize;
46 boolean _strict_file_io;
47
48 public PerVolumeJSON(String input_dir, String whitelist_filename, String langmap_directory,
49 ArrayList<String> solr_endpoints, String output_dir, int verbosity,
50 boolean icu_tokenize, boolean strict_file_io)
51 {
52 System.out.println("*** PerVolumeJSON Constructor, langmap_directory = " + langmap_directory);
53
54 _input_dir = input_dir;
55 _whitelist_filename = whitelist_filename;
56 _langmap_directory = langmap_directory;
57
58 _solr_endpoints = solr_endpoints;
59 _solr_endpoints_len = solr_endpoints.size();
60
61 //_solr_url = solr_url;
62 _output_dir = output_dir;
63 _verbosity = verbosity;
64
65 _icu_tokenize = icu_tokenize;
66 _strict_file_io = strict_file_io;
67
68 _whitelist_bloomfilter = null;
69 _universal_langmap = null;
70 }
71
72
73 public Integer call(Text json_text) throws IOException
74
75 {
76 if (_whitelist_filename != null) {
77
78 synchronized (_whitelist_filename) {
79 if (_whitelist_bloomfilter == null) {
80
81 _whitelist_bloomfilter = new WhitelistBloomFilter(_whitelist_filename,true);
82 }
83 }
84 }
85
86 if (_langmap_directory != null) {
87
88 synchronized (_langmap_directory) {
89 if (_universal_langmap == null) {
90 _universal_langmap = new UniversalPOSLangMap(_langmap_directory);
91 }
92 }
93 }
94
95 int ef_num_pages = 0;
96
97 String solr_url = null;
98 if (_solr_endpoints_len > 0) {
99 int random_choice = (int)(_solr_endpoints_len * Math.random());
100 solr_url = _solr_endpoints.get(random_choice);
101 }
102
103 try {
104
105
106 JSONObject extracted_feature_record = new JSONObject(json_text.toString());
107
108 if (extracted_feature_record != null) {
109 String volume_id = extracted_feature_record.getString("id");
110
111 JSONObject ef_metadata = extracted_feature_record.getJSONObject("metadata");
112 //String title= ef_metadata.getString("title");
113
114 //
115 // Top-level metadata Solr doc
116 //
117 JSONObject solr_add_metadata_doc_json = SolrDocJSON.generateToplevelMetadataSolrDocJSON(volume_id,ef_metadata);
118 if (solr_add_metadata_doc_json != null) {
119
120 if ((_verbosity >=2)) {
121 System.out.println("==================");
122 System.out.println("Metadata JSON: " + solr_add_metadata_doc_json.toString());
123 System.out.println("==================");
124 }
125
126 if (solr_url != null) {
127
128 if ((_verbosity >=2) ) {
129 System.out.println("==================");
130 System.out.println("Posting to: " + solr_url);
131 System.out.println("==================");
132 }
133 SolrDocJSON.postSolrDoc(solr_url, solr_add_metadata_doc_json, volume_id, "top-level-metadata");
134 }
135 }
136
137 //
138 // Now move on to POS extracted features per-page
139 //
140 JSONObject ef_features = extracted_feature_record.getJSONObject("features");
141
142 int ef_page_count = ef_features.getInt("pageCount");
143
144 if (_verbosity >= 1) {
145 System.out.println("Processing: " + volume_id);
146 System.out.println(" pageCount = " + ef_page_count);
147 }
148
149 JSONArray ef_pages = ef_features.getJSONArray("pages");
150 ef_num_pages = ef_pages.length();
151
152
153 for (int i = 0; i < ef_page_count; i++) {
154 String formatted_i = String.format("page-%06d", i);
155 String page_id = volume_id + "." + formatted_i;
156
157 if (_verbosity >= 2) {
158 System.out.println(" Page: " + page_id);
159 }
160
161
162 JSONObject ef_page = ef_pages.getJSONObject(i);
163
164 if (ef_page != null) {
165 // Convert to Solr add form
166 JSONObject solr_add_doc_json
167 = SolrDocJSON.generateSolrDocJSON(volume_id, page_id, ef_page, _whitelist_bloomfilter, _universal_langmap, _icu_tokenize);
168
169
170 if ((_verbosity >=2) && (i==20)) {
171 System.out.println("==================");
172 System.out.println("Sample output Solr add JSON [page 20]: " + solr_add_doc_json.toString());
173 System.out.println("==================");
174 }
175
176
177 if (solr_url != null) {
178 SolrDocJSON.postSolrDoc(solr_url, solr_add_doc_json,
179 volume_id, page_id);
180 }
181 }
182 else {
183 System.err.println("Skipping: " + page_id);
184 }
185
186 }
187 }
188 }
189 catch (Exception e) {
190 if (_strict_file_io) {
191 throw e;
192 }
193 else {
194 e.printStackTrace();
195 }
196 }
197
198 return ef_num_pages;
199
200 }
201
202 /*
203 //public void call(String json_file_in) throws IOException
204 public Integer call(String json_file_in) throws IOException
205
206 {
207 if ((_whitelist_filename != null) && (_whitelist_bloomfilter == null)) {
208 _whitelist_bloomfilter = new WhitelistBloomFilter(_whitelist_filename,true);
209 }
210
211 int ef_num_pages = 0;
212
213 ArrayList<String> ids = new ArrayList<String>(); // want it to be non-null so can return valid iterator
214
215 String full_json_file_in = _input_dir + "/" + json_file_in;
216 JSONObject extracted_feature_record = JSONClusterFileIO.readJSONFile(full_json_file_in);
217
218 if (extracted_feature_record != null) {
219 String volume_id = extracted_feature_record.getString("id");
220
221 //JSONObject ef_metadata = extracted_feature_record.getJSONObject("metadata");
222 //String title= ef_metadata.getString("title");
223
224 JSONObject ef_features = extracted_feature_record.getJSONObject("features");
225
226 int ef_page_count = ef_features.getInt("pageCount");
227
228 if (_verbosity >= 1) {
229 System.out.println("Processing: " + json_file_in);
230 System.out.println(" pageCount = " + ef_page_count);
231 }
232
233 JSONArray ef_pages = ef_features.getJSONArray("pages");
234 ef_num_pages = ef_pages.length();
235
236 // Make directory for page-level JSON output
237 String json_dir = ClusterFileIO.removeSuffix(json_file_in,".json.bz2");
238 String page_json_dir = json_dir + "/pages";
239
240 if (_output_dir != null) {
241 ClusterFileIO.createDirectoryAll(_output_dir + "/" + page_json_dir);
242 }
243
244 ids = new ArrayList<String>(ef_num_pages);
245 for (int i = 0; i < ef_page_count; i++) {
246 String formatted_i = String.format("page-%06d", i);
247 String page_id = volume_id + "." + formatted_i;
248
249 if (_verbosity >= 2) {
250 System.out.println(" Page: " + page_id);
251 }
252
253 String output_json_bz2 = page_json_dir +"/" + formatted_i + ".json.bz2";
254 ids.add(page_id);
255
256 if (_verbosity >=2) {
257 if (i==0) {
258 System.out.println("Sample output JSON page file [i=0]: " + output_json_bz2);
259 }
260 }
261 JSONObject ef_page = ef_pages.getJSONObject(i);
262
263 if (ef_page != null) {
264 // Convert to Solr add form
265 JSONObject solr_add_doc_json
266 = SolrDocJSON.generateSolrDocJSON(volume_id, page_id, ef_page, _whitelist_bloomfilter, _icu_tokenize);
267
268
269 if ((_verbosity >=2) && (i==20)) {
270 System.out.println("==================");
271 System.out.println("Sample output Solr add JSON [page 20]: " + solr_add_doc_json.toString());
272 System.out.println("==================");
273 }
274
275
276 if (_solr_url != null) {
277 if ((_verbosity >=2) && (i==20)) {
278 System.out.println("==================");
279 System.out.println("Posting to: " + _solr_url);
280 System.out.println("==================");
281 }
282 SolrDocJSON.postSolrDoc(_solr_url, solr_add_doc_json);
283 }
284
285 if (_output_dir != null) {
286 if ((_verbosity >=2) && (i==20)) {
287 System.out.println("==================");
288 System.out.println("Saving to: " + _output_dir);
289 System.out.println("==================");
290 }
291 SolrDocJSON.saveSolrDoc(solr_add_doc_json, _output_dir + "/" + output_json_bz2);
292 }
293 }
294 else {
295 System.err.println("Skipping: " + page_id);
296 }
297
298 }
299 }
300 else {
301 // File did not exist, or could not be parsed
302 String mess = "Failed to read in bzipped JSON file '" + full_json_file_in + "'";
303 if (_strict_file_io) {
304 throw new IOException(mess);
305 }
306 else {
307 System.err.println("Warning: " + mess);
308 System.out.println("Warning: " + mess);
309 }
310 }
311
312 return ef_num_pages;
313
314 }
315 */
316}
317
Note: See TracBrowser for help on using the repository browser.