Java tutorial
/* * Copyright 2012-2016 CodeLibs Project and the Others. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, * either express or implied. See the License for the specific language * governing permissions and limitations under the License. */ package org.codelibs.fess.crawler.extractor.impl; import java.io.InputStream; import java.util.HashMap; import java.util.Map; import javax.annotation.Resource; import org.apache.commons.compress.archivers.ArchiveInputStream; import org.apache.commons.compress.archivers.ArchiveStreamFactory; import org.apache.commons.compress.archivers.zip.ZipArchiveEntry; import org.apache.commons.io.IOUtils; import org.apache.tika.metadata.TikaMetadataKeys; import org.codelibs.fess.crawler.container.CrawlerContainer; import org.codelibs.fess.crawler.entity.ExtractData; import org.codelibs.fess.crawler.exception.CrawlerSystemException; import org.codelibs.fess.crawler.exception.ExtractException; import org.codelibs.fess.crawler.exception.MaxLengthExceededException; import org.codelibs.fess.crawler.extractor.Extractor; import org.codelibs.fess.crawler.extractor.ExtractorFactory; import org.codelibs.fess.crawler.helper.MimeTypeHelper; import org.codelibs.fess.crawler.util.IgnoreCloseInputStream; import org.codelibs.fess.crawler.util.UnsafeStringBuilder; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * @author shinsuke * */ public class ZipExtractor implements Extractor { private static final Logger logger = LoggerFactory.getLogger(ZipExtractor.class); @Resource protected CrawlerContainer crawlerContainer; @Resource protected ArchiveStreamFactory archiveStreamFactory; protected long maxContentSize = -1; @Override public ExtractData getText(final InputStream in, final Map<String, String> params) { if (in == null) { throw new CrawlerSystemException("The inputstream is null."); } final MimeTypeHelper mimeTypeHelper = crawlerContainer.getComponent("mimeTypeHelper"); if (mimeTypeHelper == null) { throw new CrawlerSystemException("MimeTypeHelper is unavailable."); } final ExtractorFactory extractorFactory = crawlerContainer.getComponent("extractorFactory"); if (extractorFactory == null) { throw new CrawlerSystemException("ExtractorFactory is unavailable."); } final UnsafeStringBuilder buf = new UnsafeStringBuilder(1000); ArchiveInputStream ais = null; try { ais = archiveStreamFactory.createArchiveInputStream(in); ZipArchiveEntry entry = null; long contentSize = 0; while ((entry = (ZipArchiveEntry) ais.getNextEntry()) != null) { contentSize += entry.getSize(); if (maxContentSize != -1 && contentSize > maxContentSize) { throw new MaxLengthExceededException( "Extracted size is " + contentSize + " > " + maxContentSize); } final String filename = entry.getName(); final String mimeType = mimeTypeHelper.getContentType(null, filename); if (mimeType != null) { final Extractor extractor = extractorFactory.getExtractor(mimeType); if (extractor != null) { try { final Map<String, String> map = new HashMap<String, String>(); map.put(TikaMetadataKeys.RESOURCE_NAME_KEY, filename); buf.append(extractor.getText(new IgnoreCloseInputStream(ais), map).getContent()); buf.append('\n'); } catch (final Exception e) { if (logger.isDebugEnabled()) { logger.debug("Exception in an internal extractor.", e); } } } } } } catch (final MaxLengthExceededException e) { throw e; } catch (final Exception e) { if (buf.length() == 0) { throw new ExtractException("Could not extract a content.", e); } } finally { IOUtils.closeQuietly(ais); } return new ExtractData(buf.toUnsafeString().trim()); } public void setMaxContentSize(long maxContentSize) { this.maxContentSize = maxContentSize; } }