org.toubassi.femtozip.lucene.StoredFieldDumper.java Source code

Java tutorial

Introduction

Here is the source code for org.toubassi.femtozip.lucene.StoredFieldDumper.java

Source

/**
 *   Copyright 2011 Garrick Toubassi
 *
 *   Licensed under the Apache License, Version 2.0 (the "License");
 *   you may not use this file except in compliance with the License.
 *   You may obtain a copy of the License at
 *
 *       http://www.apache.org/licenses/LICENSE-2.0
 *
 *   Unless required by applicable law or agreed to in writing, software
 *   distributed under the License is distributed on an "AS IS" BASIS,
 *   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 *   See the License for the specific language governing permissions and
 *   limitations under the License.
 */
package org.toubassi.femtozip.lucene;

import java.io.BufferedOutputStream;
import java.io.FileOutputStream;
import java.io.IOException;
import java.io.OutputStream;
import java.util.Collection;
import java.util.HashMap;
import java.util.Map;

import org.apache.lucene.document.Document;
import org.apache.lucene.document.Field;
import org.apache.lucene.index.IndexReader;

public class StoredFieldDumper {

    private String indexPath;
    private String outputBasePath;

    protected void parseArguments(String[] args) {

        for (int i = 0, count = args.length; i < count; i++) {
            String arg = args[i];

            if (arg.equals("--outputbasepath")) {
                outputBasePath = args[++i];
            } else {
                indexPath = arg;
            }
        }
    }

    protected void dump() throws IOException {
        IndexReader reader = IndexReader.open(indexPath);

        Collection<?> allFields = reader.getFieldNames(IndexReader.FieldOption.ALL);
        String[] fieldNames = new String[allFields.size()];
        allFields.toArray(fieldNames);

        Map<String, OutputStream> output = new HashMap<String, OutputStream>();

        long lastStatusTime = 0;

        for (int docId = 0, count = reader.maxDoc(); docId < count; docId++) {
            Document doc = reader.document(docId);

            if (System.currentTimeMillis() - lastStatusTime > 5000) {
                lastStatusTime = System.currentTimeMillis();
                System.out.println("Processing docId " + docId + " of " + count);
            }

            for (String fieldName : fieldNames) {
                Field[] fields = doc.getFields(fieldName);

                for (Field field : fields) {

                    if (!field.isStored() || field.isCompressed()) {
                        // TODO if its compressed, uncompress it and benchmark it.
                        continue;
                    }

                    byte[] bytes;

                    if (field.isBinary()) {
                        bytes = new byte[field.getBinaryLength()];
                        System.arraycopy(field.getBinaryValue(), field.getBinaryOffset(), bytes, 0,
                                field.getBinaryLength());
                    } else {
                        String value = field.stringValue();
                        bytes = value.getBytes("UTF-8");
                    }

                    OutputStream out = output.get(fieldName);
                    if (out == null) {
                        FileOutputStream fileOut = new FileOutputStream(outputBasePath + "_" + fieldName);
                        out = new BufferedOutputStream(fileOut);
                        output.put(fieldName, out);
                    }

                    out.write(bytes);
                }
            }
        }

        reader.close();

        for (Map.Entry<String, OutputStream> entry : output.entrySet()) {
            entry.getValue().close();
        }
    }

    protected void run(String[] args) throws IOException {
        parseArguments(args);

        dump();
    }

    public static void main(String[] args) throws IOException {
        StoredFieldDumper dumper = new StoredFieldDumper();
        dumper.run(args);
    }
}