List of usage examples for com.google.common.collect Lists newArrayList
@GwtCompatible(serializable = true) public static <E> ArrayList<E> newArrayList()
From source file:org.apache.shindig.gadgets.oauth.OAuthCommandLine.java
public static void main(String[] argv) throws Exception { Map<String, String> params = Maps.newHashMap(); for (int i = 0; i < argv.length; i += 2) { params.put(argv[i], argv[i + 1]); }//from w ww .java2 s .co m final String httpProxy = params.get("--httpProxy"); final String consumerKey = params.get("--consumerKey"); final String consumerSecret = params.get("--consumerSecret"); final String xOauthRequestor = params.get("--requestorId"); final String accessToken = params.get("--accessToken"); final String tokenSecret = params.get("--tokenSecret"); final String method = params.get("--method") == null ? "GET" : params.get("--method"); String url = params.get("--url"); String contentType = params.get("--contentType"); String postBody = params.get("--postBody"); String postFile = params.get("--postFile"); String paramLocation = params.get("--paramLocation"); String bodySigning = params.get("--bodySigning"); HttpRequest request = new HttpRequest(Uri.parse(url)); if (contentType != null) { request.setHeader("Content-Type", contentType); } else { request.setHeader("Content-Type", OAuth.FORM_ENCODED); } if (postBody != null) { request.setPostBody(postBody.getBytes()); } if (postFile != null) { request.setPostBody(IOUtils.toByteArray(new FileInputStream(postFile))); } OAuthParamLocation paramLocationEnum = OAuthParamLocation.URI_QUERY; if (paramLocation != null) { paramLocationEnum = OAuthParamLocation.valueOf(paramLocation); } BodySigning bodySigningEnum = BodySigning.none; if (bodySigning != null) { bodySigningEnum = BodySigning.valueOf(bodySigning); } List<OAuth.Parameter> oauthParams = Lists.newArrayList(); UriBuilder target = new UriBuilder(Uri.parse(url)); String query = target.getQuery(); target.setQuery(null); oauthParams.addAll(OAuth.decodeForm(query)); if (OAuth.isFormEncoded(contentType) && request.getPostBodyAsString() != null) { oauthParams.addAll(OAuth.decodeForm(request.getPostBodyAsString())); } else if (bodySigningEnum == BodySigning.legacy) { oauthParams.add(new OAuth.Parameter(request.getPostBodyAsString(), "")); } else if (bodySigningEnum == BodySigning.hash) { oauthParams.add(new OAuth.Parameter(OAuthConstants.OAUTH_BODY_HASH, new String( Base64.encodeBase64(DigestUtils.sha(request.getPostBodyAsString().getBytes())), "UTF-8"))); } if (consumerKey != null) { oauthParams.add(new OAuth.Parameter(OAuth.OAUTH_CONSUMER_KEY, consumerKey)); } if (xOauthRequestor != null) { oauthParams.add(new OAuth.Parameter("xoauth_requestor_id", xOauthRequestor)); } OAuthConsumer consumer = new OAuthConsumer(null, consumerKey, consumerSecret, null); OAuthAccessor accessor = new OAuthAccessor(consumer); accessor.accessToken = accessToken; accessor.tokenSecret = tokenSecret; OAuthMessage message = accessor.newRequestMessage(method, target.toString(), oauthParams); List<Map.Entry<String, String>> entryList = OAuthRequest.selectOAuthParams(message); switch (paramLocationEnum) { case AUTH_HEADER: request.addHeader("Authorization", OAuthRequest.getAuthorizationHeader(entryList)); break; case POST_BODY: if (!OAuth.isFormEncoded(contentType)) { throw new RuntimeException("OAuth param location can only be post_body if post body if of " + "type x-www-form-urlencoded"); } String oauthData = OAuthUtil.formEncode(message.getParameters()); request.setPostBody(CharsetUtil.getUtf8Bytes(oauthData)); break; case URI_QUERY: request.setUri(Uri.parse(OAuthUtil.addParameters(request.getUri().toString(), entryList))); break; } request.setMethod(method); HttpFetcher fetcher = new BasicHttpFetcher(httpProxy); HttpResponse response = fetcher.fetch(request); System.out.println("Request ------------------------------"); System.out.println(request.toString()); System.out.println("Response -----------------------------"); System.out.println(response.toString()); }
From source file:org.apache.mahout.classifier.sgd.TrainNewsGroups.java
public static void main(String[] args) throws IOException { File base = new File(args[0]); Multiset<String> overallCounts = HashMultiset.create(); int leakType = 0; if (args.length > 1) { leakType = Integer.parseInt(args[1]); }//from w w w .ja v a 2s . com Dictionary newsGroups = new Dictionary(); NewsgroupHelper helper = new NewsgroupHelper(); helper.getEncoder().setProbes(2); AdaptiveLogisticRegression learningAlgorithm = new AdaptiveLogisticRegression(20, NewsgroupHelper.FEATURES, new L1()); learningAlgorithm.setInterval(800); learningAlgorithm.setAveragingWindow(500); List<File> files = Lists.newArrayList(); for (File newsgroup : base.listFiles()) { if (newsgroup.isDirectory()) { newsGroups.intern(newsgroup.getName()); files.addAll(Arrays.asList(newsgroup.listFiles())); } } Collections.shuffle(files); System.out.println(files.size() + " training files"); SGDInfo info = new SGDInfo(); int k = 0; for (File file : files) { String ng = file.getParentFile().getName(); int actual = newsGroups.intern(ng); Vector v = helper.encodeFeatureVector(file, actual, leakType, overallCounts); learningAlgorithm.train(actual, v); k++; State<AdaptiveLogisticRegression.Wrapper, CrossFoldLearner> best = learningAlgorithm.getBest(); SGDHelper.analyzeState(info, leakType, k, best); } learningAlgorithm.close(); SGDHelper.dissect(leakType, newsGroups, learningAlgorithm, files, overallCounts); System.out.println("exiting main"); ModelSerializer.writeBinary("/tmp/news-group.model", learningAlgorithm.getBest().getPayload().getLearner().getModels().get(0)); List<Integer> counts = Lists.newArrayList(); System.out.println("Word counts"); for (String count : overallCounts.elementSet()) { counts.add(overallCounts.count(count)); } Collections.sort(counts, Ordering.natural().reverse()); k = 0; for (Integer count : counts) { System.out.println(k + "\t" + count); k++; if (k > 1000) { break; } } }
From source file:org.apache.mahout.classifier.sequencelearning.hmm.BaumWelchTrainer.java
public static void main(String[] args) throws IOException { DefaultOptionBuilder optionBuilder = new DefaultOptionBuilder(); ArgumentBuilder argumentBuilder = new ArgumentBuilder(); Option inputOption = DefaultOptionCreator.inputOption().create(); Option outputOption = DefaultOptionCreator.outputOption().create(); Option stateNumberOption = optionBuilder.withLongName("nrOfHiddenStates") .withDescription("Number of hidden states").withShortName("nh") .withArgument(argumentBuilder.withMaximum(1).withMinimum(1).withName("number").create()) .withRequired(true).create(); Option observedStateNumberOption = optionBuilder.withLongName("nrOfObservedStates") .withDescription("Number of observed states").withShortName("no") .withArgument(argumentBuilder.withMaximum(1).withMinimum(1).withName("number").create()) .withRequired(true).create(); Option epsilonOption = optionBuilder.withLongName("epsilon").withDescription("Convergence threshold") .withShortName("e") .withArgument(argumentBuilder.withMaximum(1).withMinimum(1).withName("number").create()) .withRequired(true).create(); Option iterationsOption = optionBuilder.withLongName("max-iterations") .withDescription("Maximum iterations number").withShortName("m") .withArgument(argumentBuilder.withMaximum(1).withMinimum(1).withName("number").create()) .withRequired(true).create(); Group optionGroup = new GroupBuilder().withOption(inputOption).withOption(outputOption) .withOption(stateNumberOption).withOption(observedStateNumberOption).withOption(epsilonOption) .withOption(iterationsOption).withName("Options").create(); try {/*from w w w. jav a 2 s . co m*/ Parser parser = new Parser(); parser.setGroup(optionGroup); CommandLine commandLine = parser.parse(args); String input = (String) commandLine.getValue(inputOption); String output = (String) commandLine.getValue(outputOption); int nrOfHiddenStates = Integer.parseInt((String) commandLine.getValue(stateNumberOption)); int nrOfObservedStates = Integer.parseInt((String) commandLine.getValue(observedStateNumberOption)); double epsilon = Double.parseDouble((String) commandLine.getValue(epsilonOption)); int maxIterations = Integer.parseInt((String) commandLine.getValue(iterationsOption)); //constructing random-generated HMM HmmModel model = new HmmModel(nrOfHiddenStates, nrOfObservedStates, new Date().getTime()); List<Integer> observations = Lists.newArrayList(); //reading observations Scanner scanner = new Scanner(new FileInputStream(input), "UTF-8"); try { while (scanner.hasNextInt()) { observations.add(scanner.nextInt()); } } finally { scanner.close(); } int[] observationsArray = new int[observations.size()]; for (int i = 0; i < observations.size(); ++i) { observationsArray[i] = observations.get(i); } //training HmmModel trainedModel = HmmTrainer.trainBaumWelch(model, observationsArray, epsilon, maxIterations, true); //serializing trained model DataOutputStream stream = new DataOutputStream(new FileOutputStream(output)); try { LossyHmmSerializer.serialize(trainedModel, stream); } finally { Closeables.close(stream, false); } //printing tranied model System.out.println("Initial probabilities: "); for (int i = 0; i < trainedModel.getNrOfHiddenStates(); ++i) { System.out.print(i + " "); } System.out.println(); for (int i = 0; i < trainedModel.getNrOfHiddenStates(); ++i) { System.out.print(trainedModel.getInitialProbabilities().get(i) + " "); } System.out.println(); System.out.println("Transition matrix:"); System.out.print(" "); for (int i = 0; i < trainedModel.getNrOfHiddenStates(); ++i) { System.out.print(i + " "); } System.out.println(); for (int i = 0; i < trainedModel.getNrOfHiddenStates(); ++i) { System.out.print(i + " "); for (int j = 0; j < trainedModel.getNrOfHiddenStates(); ++j) { System.out.print(trainedModel.getTransitionMatrix().get(i, j) + " "); } System.out.println(); } System.out.println("Emission matrix: "); System.out.print(" "); for (int i = 0; i < trainedModel.getNrOfOutputStates(); ++i) { System.out.print(i + " "); } System.out.println(); for (int i = 0; i < trainedModel.getNrOfHiddenStates(); ++i) { System.out.print(i + " "); for (int j = 0; j < trainedModel.getNrOfOutputStates(); ++j) { System.out.print(trainedModel.getEmissionMatrix().get(i, j) + " "); } System.out.println(); } } catch (OptionException e) { CommandLineUtil.printHelp(optionGroup); } }
From source file:scoutdoc.main.Main.java
/** * @param args/* w ww . ja v a 2s . c o m*/ */ public static void main(String[] args) { Option optHelp = new Option(HELP_ID, "help", false, "print this message"); Option optProp = new Option(PROP_ID, "config", true, "configuration file"); optProp.setArgName("file"); //source Option optTask = new Option(SOURCE_TASKS_ID, "task", true, "(source) one or many task files"); optTask.setArgName("files"); optTask.setArgs(Option.UNLIMITED_VALUES); Option optAllPages = new Option(SOURCE_ALL_PAGES_ID, "pages", false, "(source) use the pages contained in the source folder"); Option optListPages = new Option(SOURCE_LIST_ID, "list", true, "(source) list of pages contained in the file"); optListPages.setArgName("file"); Option optRecentChange = new Option(SOURCE_RECENT_CHANGES_ID, "recent-changes", false, "(source) use the pages from the wiki recent changes"); Option optRss = new Option(SOURCE_RSS_ID, "rss", false, "(source) use the pages from the rss feed of the wiki"); OptionGroup sourceGroup = new OptionGroup(); sourceGroup.setRequired(true); sourceGroup.addOption(optTask); sourceGroup.addOption(optAllPages); sourceGroup.addOption(optListPages); sourceGroup.addOption(optRecentChange); sourceGroup.addOption(optRss); Option optfilter = new Option(SOURCE_FILTER_ID, "filter", true, "Filter for list of pages used as source"); optfilter.setArgName("class"); List<String> values = Lists.newArrayList(); for (Operation o : Operation.values()) { values.add(o.name()); } Option optOperation = new Option(OPERATION_ID, "operation", true, "operation: " + Joiner.on(", ").join(values)); optOperation.setArgName("operations"); optOperation.setArgs(Option.UNLIMITED_VALUES); optOperation.setRequired(true); Option optOutputCheckstyle = new Option(OUTPUT_CHECKSTYLE_ID, "output-checkstyle", true, "(CHECK output) create a xml checkstyle file (<filename> is optional. Default: " + DEFAULT_CHECKSTYLE_NAME + ")"); optOutputCheckstyle.setArgName("filename"); optOutputCheckstyle.setOptionalArg(true); Option optOutputDashboard = new Option(OUTPUT_DASHBOARD_ID, "output-dashboard", true, "(CHECK output) create an html dashboard (<folder> is optional. Default: " + DEFAULT_DASHBOARD_NAME + ")"); optOutputDashboard.setArgName("folder"); optOutputDashboard.setOptionalArg(true); Options options = new Options(); options.addOption(optHelp); options.addOption(optProp); options.addOptionGroup(sourceGroup); options.addOption(optfilter); options.addOption(optOperation); options.addOption(optOutputCheckstyle); options.addOption(optOutputDashboard); try { CommandLineParser parser = new GnuParser(); CommandLine cmd = parser.parse(options, args); if (cmd.hasOption(HELP_ID)) { printHelpAndExit(options); } if (cmd.hasOption(PROP_ID)) { ProjectProperties.initProperties(cmd.getOptionValue(PROP_ID)); } Pages.initPageList(); List<Operation> operations = readOptionEnum(cmd, optOperation, Operation.class); List<Task> tasks = readTasks(cmd, optTask); Collection<Page> pageList; if (cmd.hasOption(SOURCE_ALL_PAGES_ID)) { pageList = PageUtility.loadPages(ProjectProperties.getFolderWikiSource()); } else if (cmd.hasOption(SOURCE_LIST_ID)) { String name = cmd.getOptionValue(SOURCE_LIST_ID); try { pageList = PageUtility.readList(name); } catch (IOException e) { throw new MissingArgumentException("IOException for file <" + name + "> for <" + optListPages.getLongOpt() + "> : " + e.getMessage()); } } else { pageList = Collections.emptyList(); } IPageFilter filter; if (cmd.hasOption(SOURCE_FILTER_ID)) { if (tasks.size() > 0) { throw new MissingArgumentException("Filter <" + optfilter.getLongOpt() + "> is not allowed for source <" + optTask.getLongOpt() + ">."); } filter = newInstance(cmd.getOptionValue(SOURCE_FILTER_ID), IPageFilter.class, new AcceptAllPageFilter()); } else { filter = new AcceptAllPageFilter(); } List<Page> pages = Lists.newArrayList(); for (Page page : pageList) { if (filter.keepPage(page)) { pages.add(page); } } if (operations.contains(Operation.FETCH)) { if (pages.size() > 0) { ScoutDocFetch sdf = new ScoutDocFetch(); RelatedPagesStrategy strategy; if (cmd.hasOption(SOURCE_ALL_PAGES_ID)) { strategy = RelatedPagesStrategy.NO_RELATED_PAGES; } else if (cmd.hasOption(SOURCE_LIST_ID)) { strategy = RelatedPagesStrategy.IMAGES_TEMPLATES_AND_LINKS; } else { throw new IllegalStateException("Page list comes from an unexpected option"); } sdf.execute(pages, strategy); } else if (cmd.hasOption(SOURCE_RECENT_CHANGES_ID)) { ScoutDocFetch sdf = new ScoutDocFetch(); sdf.executeRecentChanges(filter); } else if (cmd.hasOption(SOURCE_RSS_ID)) { ScoutDocFetch sdf = new ScoutDocFetch(); sdf.executeRss(filter); } else if (tasks.size() > 0) { for (Task task : tasks) { ScoutDocFetch sdf = new ScoutDocFetch(); sdf.execute(task); } } else { throw new MissingArgumentException("Missing a source"); } } if (operations.contains(Operation.CHECK)) { ScoutDocCheck sdc = new ScoutDocCheck(); List<Check> checks = Lists.newArrayList(); ensureNotSet(cmd, optRecentChange, Operation.CHECK); ensureNotSet(cmd, optRss, Operation.CHECK); if (pages.size() > 0) { checks = sdc.analysePages(pages); } else if (tasks.size() > 0) { for (Task task : tasks) { ScoutDocCheck sdcForTask = new ScoutDocCheck(); checks.addAll(sdcForTask.execute(task)); } } else { throw new MissingArgumentException("Missing a source"); } //output: if (cmd.hasOption(OUTPUT_CHECKSTYLE_ID)) { String fileName = cmd.getOptionValue(OUTPUT_CHECKSTYLE_ID, DEFAULT_CHECKSTYLE_NAME); sdc.writeCheckstyleFile(checks, fileName); } if (cmd.hasOption(OUTPUT_DASHBOARD_ID)) { String folderName = cmd.getOptionValue(OUTPUT_DASHBOARD_ID, DEFAULT_DASHBOARD_NAME); sdc.writeDashboardFiles(checks, folderName); } } if (operations.contains(Operation.CONVERT)) { ensureNotSet(cmd, optAllPages, Operation.CONVERT); ensureNotSet(cmd, optListPages, Operation.CONVERT); ensureNotSet(cmd, optRecentChange, Operation.CONVERT); ensureNotSet(cmd, optRss, Operation.CONVERT); if (tasks.size() > 0) { for (Task task : tasks) { ScoutDocConverter sdc = new ScoutDocConverter(); sdc.execute(task); } } else { throw new MissingArgumentException("Missing a source"); } } } catch (MissingOptionException e) { // Check if it is an error or if optHelp was selected. boolean help = false; try { Options helpOptions = new Options(); helpOptions.addOption(optHelp); CommandLineParser parser = new PosixParser(); CommandLine line = parser.parse(helpOptions, args); if (line.hasOption(HELP_ID)) { help = true; } } catch (Exception ex) { ex.printStackTrace(); } if (!help) { System.err.println(e.getMessage()); System.err.flush(); } printHelpAndExit(options); } catch (MissingArgumentException e) { System.err.println(e.getMessage()); printHelpAndExit(options); } catch (ParseException e) { System.err.println("Error while parsing the command line: " + e.getMessage()); System.exit(1); } catch (Exception e) { e.printStackTrace(); } }
From source file:com.sina.dip.twill.HelloWorldServiceDiscovery.java
public static void main(String[] args) { String zkStr = "localhost:2181"; YarnConfiguration yarnConfiguration = new YarnConfiguration(); final TwillRunnerService twillRunner = new YarnTwillRunnerService(yarnConfiguration, zkStr); twillRunner.start();//from ww w.j av a2 s. c om String yarnClasspath = yarnConfiguration.get(YarnConfiguration.YARN_APPLICATION_CLASSPATH, "/usr/lib/hadoop/*,/usr/lib/hadoop-0.20-mapreduce/*,/usr/lib/hadoop-hdfs/*,/usr/lib/hadoop-mapreduce/*,/usr/lib/hadoop-yarn/*"); List<String> applicationClassPaths = Lists.newArrayList(); Iterables.addAll(applicationClassPaths, Splitter.on(",").split(yarnClasspath)); final TwillController controller = twillRunner.prepare(new HelloWorldApplication()) .addLogHandler(new PrinterLogHandler(new PrintWriter(System.out, true))) .withApplicationClassPaths(applicationClassPaths) .withBundlerClassAcceptor(new HadoopClassExcluder()).start(); ServiceDiscovered helloWorldService = controller.discoverService("HelloWorldService"); ServiceDiscovered helloWorldService2 = controller.discoverService("HelloWorldService2"); int count = 0; while (true) { boolean flag = true; Iterator<Discoverable> iterator = helloWorldService.iterator(); while (iterator.hasNext()) { Discoverable discoverable = iterator.next(); System.out.println(discoverable.getName() + " : " + discoverable.getSocketAddress()); flag = false; } iterator = helloWorldService2.iterator(); while (iterator.hasNext()) { Discoverable discoverable = iterator.next(); System.out.println(discoverable.getName() + " : " + discoverable.getSocketAddress()); flag = false; } try { Thread.sleep(5 * 1000); } catch (InterruptedException e) { } if (++count >= 36 && flag) { break; } } Runtime.getRuntime().addShutdownHook(new Thread() { @Override public void run() { try { Futures.getUnchecked(controller.terminate()); } finally { twillRunner.stop(); } } }); try { controller.awaitTerminated(); } catch (ExecutionException e) { e.printStackTrace(); } }
From source file:com.sina.dip.twill.HelloWorldControllingLiveApplications.java
public static void main(String[] args) { String zkStr = "localhost:2181"; YarnConfiguration yarnConfiguration = new YarnConfiguration(); final TwillRunnerService twillRunner = new YarnTwillRunnerService(yarnConfiguration, zkStr); twillRunner.start();// www . j a v a 2s. c o m String yarnClasspath = yarnConfiguration.get(YarnConfiguration.YARN_APPLICATION_CLASSPATH, "/usr/lib/hadoop/*,/usr/lib/hadoop-0.20-mapreduce/*,/usr/lib/hadoop-hdfs/*,/usr/lib/hadoop-mapreduce/*,/usr/lib/hadoop-yarn/*"); List<String> applicationClassPaths = Lists.newArrayList(); Iterables.addAll(applicationClassPaths, Splitter.on(",").split(yarnClasspath)); final TwillController controller = twillRunner.prepare(new HelloWorldApplication()) .addLogHandler(new PrinterLogHandler(new PrintWriter(System.out, true))) .withApplicationClassPaths(applicationClassPaths) .withBundlerClassAcceptor(new HadoopClassExcluder()).start(); ServiceDiscovered helloWorldService = controller.discoverService("HelloWorldService"); ServiceDiscovered helloWorldService2 = controller.discoverService("HelloWorldService2"); int count = 0; while (true) { boolean flag = true; Iterator<Discoverable> iterator = helloWorldService.iterator(); while (iterator.hasNext()) { Discoverable discoverable = iterator.next(); System.out.println(discoverable.getName() + " : " + discoverable.getSocketAddress()); flag = false; } iterator = helloWorldService2.iterator(); while (iterator.hasNext()) { Discoverable discoverable = iterator.next(); System.out.println(discoverable.getName() + " : " + discoverable.getSocketAddress()); flag = false; } try { Thread.sleep(5 * 1000); } catch (InterruptedException e) { } ++count; if (count == 10) { controller.changeInstances("hello1", 3); controller.changeInstances("hello2", 5); } else if (count == 20) { controller.changeInstances("hello1", 5); controller.changeInstances("hello2", 3); } if (count >= 36 && flag) { break; } } Runtime.getRuntime().addShutdownHook(new Thread() { @Override public void run() { try { Futures.getUnchecked(controller.terminate()); } finally { twillRunner.stop(); } } }); try { controller.awaitTerminated(); } catch (ExecutionException e) { e.printStackTrace(); } }
From source file:natlab.Main.java
/** * Main method deals with command line options and execution of * desired functions.//from w ww.j a va 2 s . com */ public static void main(String[] args) throws Exception { if (args.length == 0) { System.err.println("No options given\nTry -help for usage"); return; } options = new Options(); options.parse(args); if (options.help()) { System.err.println(options.getUsage()); return; } if (options.pref()) { NatlabPreferences.modify(options); return; } if (options.show_pref()) { System.out.println("Preferences:"); System.out.println( Joiner.on('\n').withKeyValueSeparator(" = ").join(NatlabPreferences.getAllPreferences())); } if (options.version()) { System.out.println("The version of this release is: " + VersionInfo.getVersion()); return; } if (options.quiet()) { AbstractDepthFirstAnalysis.DEBUG = false; AbstractStructuralAnalysis.DEBUG = false; } if (options.tamer()) { //TODO - the parsing of the options should probably not be done by the tamer tool BasicTamerTool.main(options); return; } if (options.tamerplus()) { TamerPlusMain.main(options); } if (options.mclint()) { McLint.main(options); return; } if (options.server()) { NatlabServer.create(options).start(); return; } if (options.mix10c()) { Mix10.compile(options); } // Mc2For options if (options.codegen() || options.nocheck()) { Main_readable.compile(options); } if (options.getFiles().isEmpty()) { System.err.println("No files provided, must have at least one file."); return; } List<String> files = options.getFiles(); log("Parsing " + Joiner.on(", ").join(files)); List<CompilationProblem> errors = Lists.newArrayList(); CompilationUnits cu; if (!options.natlab()) { cu = Parse.parseMatlabFiles(files, errors); } else { cu = Parse.parseNatlabFiles(files, errors); } if (!errors.isEmpty()) { System.err.println(CompilationProblem.toStringAll(errors)); return; } if (options.simplify()) { Simplifier.simplify(cu, FullSimplification.class); } if (options.json()) { System.out.println(cu.getJsonString()); return; } if (options.xml()) { System.out.print(cu.XMLtoString(cu.ASTtoXML())); return; } if (options.pretty()) { log("Pretty printing"); if (options.od().length() == 0) { System.out.println(cu.getPrettyPrinted()); } else { File outputDir = new File(options.od()); for (Program program : cu.getPrograms()) { File outputFile = new File(outputDir, program.getFile().getName()); Files.createParentDirs(outputFile); Files.write(program.getPrettyPrinted(), outputFile, Charsets.UTF_8); } } return; } }
From source file:com.memonews.mahout.sentiment.SentimentModelTrainer.java
public static void main(final String[] args) throws IOException { final File base = new File(args[0]); final String modelPath = args.length > 1 ? args[1] : "target/model"; final Multiset<String> overallCounts = HashMultiset.create(); final Dictionary newsGroups = new Dictionary(); final SentimentModelHelper helper = new SentimentModelHelper(); helper.getEncoder().setProbes(2);//w w w . ja v a2 s .c o m final AdaptiveLogisticRegression learningAlgorithm = new AdaptiveLogisticRegression(2, SentimentModelHelper.FEATURES, new L1()); learningAlgorithm.setInterval(800); learningAlgorithm.setAveragingWindow(500); final List<File> files = Lists.newArrayList(); for (final File newsgroup : base.listFiles()) { if (newsgroup.isDirectory()) { newsGroups.intern(newsgroup.getName()); files.addAll(Arrays.asList(newsgroup.listFiles())); } } Collections.shuffle(files); System.out.printf("%d training files\n", files.size()); final SGDInfo info = new SGDInfo(); int k = 0; for (final File file : files) { final String ng = file.getParentFile().getName(); final int actual = newsGroups.intern(ng); final Vector v = helper.encodeFeatureVector(file, overallCounts); learningAlgorithm.train(actual, v); k++; final State<AdaptiveLogisticRegression.Wrapper, CrossFoldLearner> best = learningAlgorithm.getBest(); SGDHelper.analyzeState(info, 0, k, best); } learningAlgorithm.close(); SGDHelper.dissect(0, newsGroups, learningAlgorithm, files, overallCounts); System.out.println("exiting main"); ModelSerializer.writeBinary(modelPath, learningAlgorithm.getBest().getPayload().getLearner().getModels().get(0)); final List<Integer> counts = Lists.newArrayList(); System.out.printf("Word counts\n"); for (final String count : overallCounts.elementSet()) { counts.add(overallCounts.count(count)); } Collections.sort(counts, Ordering.natural().reverse()); k = 0; for (final Integer count : counts) { System.out.printf("%d\t%d\n", k, count); k++; if (k > 1000) { break; } } }
From source file:org.apache.ctakes.temporal.eval.EvaluationOfTimeSpans.java
public static void main(String[] args) throws Exception { Options options = CliFactory.parseArguments(Options.class, args); List<Integer> trainItems = null; List<Integer> devItems = null; List<Integer> testItems = null; List<Integer> patientSets = options.getPatients().getList(); if (options.getXMLFormat() == XMLFormat.I2B2) { trainItems = I2B2Data.getTrainPatientSets(options.getXMLDirectory()); devItems = I2B2Data.getDevPatientSets(options.getXMLDirectory()); testItems = I2B2Data.getTestPatientSets(options.getXMLDirectory()); } else {//from w w w .j a va 2s .c om trainItems = THYMEData.getPatientSets(patientSets, options.getTrainRemainders().getList()); devItems = THYMEData.getPatientSets(patientSets, options.getDevRemainders().getList()); testItems = THYMEData.getPatientSets(patientSets, options.getTestRemainders().getList()); } List<Integer> allTrain = new ArrayList<>(trainItems); List<Integer> allTest = null; if (options.getTest()) { allTrain.addAll(devItems); allTest = new ArrayList<>(testItems); } else { allTest = new ArrayList<>(devItems); } // specify the annotator classes to use List<Class<? extends JCasAnnotator_ImplBase>> annotatorClasses = Lists.newArrayList(); if (options.getRunBackwards()) annotatorClasses.add(BackwardsTimeAnnotator.class); if (options.getRunForwards()) annotatorClasses.add(TimeAnnotator.class); if (options.getRunParserBased()) annotatorClasses.add(ConstituencyBasedTimeAnnotator.class); if (options.getRunCrfBased()) annotatorClasses.add(CRFTimeAnnotator.class); if (annotatorClasses.size() == 0) { // run all annotatorClasses.add(BackwardsTimeAnnotator.class); annotatorClasses.add(TimeAnnotator.class); annotatorClasses.add(ConstituencyBasedTimeAnnotator.class); annotatorClasses.add(CRFTimeAnnotator.class); } Map<Class<? extends JCasAnnotator_ImplBase>, String[]> annotatorTrainingArguments = Maps.newHashMap(); // THYME best params: Backwards: 0.1, CRF 0.3, Time 0.1, Constituency 0.3 // i2b2 best params: Backwards 0.1, CRF 3.0, Time 0.1, Constituency 0.3 // String gridParam = "0.01"; annotatorTrainingArguments.put(BackwardsTimeAnnotator.class, new String[] { "-c", "0.1" }); annotatorTrainingArguments.put(TimeAnnotator.class, new String[] { "-c", "0.1" }); annotatorTrainingArguments.put(ConstituencyBasedTimeAnnotator.class, new String[] { "-c", "0.3" }); annotatorTrainingArguments.put(CRFTimeAnnotator.class, new String[] { "-p", "c2=" + "0.3" }); // run one evaluation per annotator class final Map<Class<?>, AnnotationStatistics<?>> annotatorStats = Maps.newHashMap(); for (Class<? extends JCasAnnotator_ImplBase> annotatorClass : annotatorClasses) { EvaluationOfTimeSpans evaluation = new EvaluationOfTimeSpans(new File("target/eval/time-spans"), options.getRawTextDirectory(), options.getXMLDirectory(), options.getXMLFormat(), options.getSubcorpus(), options.getXMIDirectory(), options.getTreebankDirectory(), options.getFeatureSelectionThreshold(), options.getSMOTENeighborNumber(), annotatorClass, options.getPrintOverlappingSpans(), annotatorTrainingArguments.get(annotatorClass)); evaluation.prepareXMIsFor(patientSets); evaluation.setSkipTrain(options.getSkipTrain()); evaluation.printErrors = options.getPrintErrors(); if (options.getI2B2Output() != null) evaluation.setI2B2Output(options.getI2B2Output() + "/" + annotatorClass.getSimpleName()); String name = String.format("%s.errors", annotatorClass.getSimpleName()); evaluation.setLogging(Level.FINE, new File("target/eval", name)); AnnotationStatistics<String> stats = evaluation.trainAndTest(allTrain, allTest); annotatorStats.put(annotatorClass, stats); } // allow ordering of models by F1 Ordering<Class<? extends JCasAnnotator_ImplBase>> byF1 = Ordering.natural() .onResultOf(new Function<Class<? extends JCasAnnotator_ImplBase>, Double>() { @Override public Double apply(Class<? extends JCasAnnotator_ImplBase> annotatorClass) { return annotatorStats.get(annotatorClass).f1(); } }); // print out models, ordered by F1 for (Class<?> annotatorClass : byF1.sortedCopy(annotatorClasses)) { System.err.printf("===== %s =====\n", annotatorClass.getSimpleName()); System.err.println(annotatorStats.get(annotatorClass)); } }
From source file:backtype.storm.command.gray_upgrade.java
public static void main(String[] args) throws Exception { if (args == null || args.length < 1) { System.out.println("Invalid parameter"); usage();//from w w w . j a va 2 s . co m return; } String topologyName = args[0]; String[] str2 = Arrays.copyOfRange(args, 1, args.length); CommandLineParser parser = new GnuParser(); Options r = buildGeneralOptions(new Options()); CommandLine commandLine = parser.parse(r, str2, true); int workerNum = 0; String component = null; List<String> workers = null; if (commandLine.hasOption("n")) { workerNum = Integer.valueOf(commandLine.getOptionValue("n")); } if (commandLine.hasOption("p")) { component = commandLine.getOptionValue("p"); } if (commandLine.hasOption("w")) { String w = commandLine.getOptionValue("w"); if (!StringUtils.isBlank(w)) { workers = Lists.newArrayList(); String[] parts = w.split(","); for (String part : parts) { if (part.split(":").length == 2) { workers.add(part.trim()); } } } } upgradeTopology(topologyName, component, workers, workerNum); }