当前位置: 首页>>代码示例>>Java>>正文


Java SparkConf.set方法代码示例

本文整理汇总了Java中org.apache.spark.SparkConf.set方法的典型用法代码示例。如果您正苦于以下问题:Java SparkConf.set方法的具体用法?Java SparkConf.set怎么用?Java SparkConf.set使用的例子?那么恭喜您, 这里精选的方法代码示例或许可以为您提供帮助。您也可以进一步了解该方法所在org.apache.spark.SparkConf的用法示例。


在下文中一共展示了SparkConf.set方法的15个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Java代码示例。

示例1: SparkDriver

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
public SparkDriver(Properties props) {
  SparkConf conf = new SparkConf().setAppName(props.getProperty(MudrodConstants.SPARK_APP_NAME, "MudrodSparkApp")).setIfMissing("spark.master", props.getProperty(MudrodConstants.SPARK_MASTER))
      .set("spark.hadoop.validateOutputSpecs", "false").set("spark.files.overwrite", "true");

  String esHost = props.getProperty(MudrodConstants.ES_UNICAST_HOSTS);
  String esPort = props.getProperty(MudrodConstants.ES_HTTP_PORT);

  if (!"".equals(esHost)) {
    conf.set("es.nodes", esHost);
  }

  if (!"".equals(esPort)) {
    conf.set("es.port", esPort);
  }

  conf.set("spark.serializer", KryoSerializer.class.getName());
  conf.set("es.batch.size.entries", "1500");

  sc = new JavaSparkContext(conf);
  sqlContext = new SQLContext(sc);
}
 
开发者ID:apache,项目名称:incubator-sdap-mudrod,代码行数:22,代码来源:SparkDriver.java

示例2: SparkTrainWorker

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
public SparkTrainWorker(
                        SparkConf conf,
                        String modelName,
                        String configPath,
                        String configFile,
                        String pyTransformScript,
                        boolean needPyTransform,
                        String loginName,
                        String hostName,
                        int hostPort,
                        int slaveNum,
                        int threadNum) throws Exception {
    super(modelName, configPath, configFile, pyTransformScript, needPyTransform,
            loginName, hostName, hostPort, threadNum);
    this.slaveNum = slaveNum;

    conf.set("spark.files.fetchTimeout", "3200");
    conf.set("spark.network.timeout", "3200");
    conf.set("spark.dynamicAllocation.executorIdleTimeout", "3200");
    conf.set("spark.dynamicAllocation.schedulerBacklogTimeout", "300");
    conf.set("spark.core.connection.auth.wait.timeout", "3200");
    conf.set("spark.memory.fraction", "0.01");
}
 
开发者ID:yuantiku,项目名称:ytk-learn,代码行数:24,代码来源:SparkTrainWorker.java

示例3: configureSparkContext

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
private void configureSparkContext(Properties properties) {
    SparkConf sparkConf = new SparkConf();
    sparkConf.setAppName("Write pipeline");
    sparkConf.set("spark.driver.allowMultipleContexts", "true");

    sparkConf.setMaster(properties.getProperty("spark.master"));
    sparkConf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer");

    sparkConf.set("spark.cassandra.connection.host", properties.getProperty("cassandra.nodes"));
    sparkConf.set("spark.cassandra.output.batch.size.bytes", properties.getProperty("cassandra.batch.size.bytes"));
    sparkConf.set("spark.cassandra.connection.port", properties.getProperty("cassandra.port"));

    sparkConf.set("es.nodes", properties.getProperty("elasticsearch.nodes") + ":" + properties.getProperty("elasticsearch.port.rest"));
    sparkConf.set("es.batch.size.entries", properties.getProperty("elasticsearch.batch.size.entries"));
    sparkConf.set("es.batch.size.bytes", properties.getProperty("elasticsearch.batch.size.bytes"));
    sparkConf.set("es.nodes.discovery", properties.getProperty("elasticsearch.nodes.dicovery"));

    sparkContext = new JavaSparkContext(sparkConf);
}
 
开发者ID:echauchot,项目名称:bigDataRocks,代码行数:20,代码来源:WritePipeline.java

示例4: getContext

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
public synchronized JavaSparkContext getContext() {
    if (sc != null)
        return sc;

    SparkConf sparkConf = new SparkConf().setMaster("local[*]").set("spark.driver.host", "localhost")
                    .set("spark.driverEnv.SPARK_LOCAL_IP", "127.0.0.1")
                    .set("spark.executorEnv.SPARK_LOCAL_IP", "127.0.0.1").setAppName("sparktest");
    if (useKryo()) {
        sparkConf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer");
    }


    sc = new JavaSparkContext(sparkConf);

    return sc;
}
 
开发者ID:deeplearning4j,项目名称:DataVec,代码行数:17,代码来源:BaseSparkTest.java

示例5: jsc

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
@Memoized
JavaStreamingContext jsc() {
  SparkConf conf = new SparkConf(true)
      .setMaster(master())
      .setAppName(getClass().getName());
  if (!jars().isEmpty()) conf.setJars(jars().toArray(new String[0]));
  for (Map.Entry<String, String> entry : conf().entrySet()) {
    conf.set(entry.getKey(), entry.getValue());
  }
  return new JavaStreamingContext(conf, new Duration(batchDuration()));
}
 
开发者ID:openzipkin,项目名称:zipkin-sparkstreaming,代码行数:12,代码来源:SparkStreamingJob.java

示例6: getSparkContext

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
public static JavaSparkContext getSparkContext(String appName){
    if(jscSingleton == null){
        SparkConf sparkConf = new SparkConf().setAppName(appName);
        sparkConf.setMaster("local[*]");
        sparkConf.set("spark.driver.maxResultSize", "16g");
        jscSingleton = new JavaSparkContext(sparkConf);
    }
    return jscSingleton;
}
 
开发者ID:tudorv91,项目名称:SparkJNI,代码行数:10,代码来源:ExampleUtils.java

示例7: getSparkConf

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
public SparkConf getSparkConf() {
    SparkConf sparkConf = new SparkConf();
    sparkConf.set("spark.streaming.kafka.maxRatePerPartition",
            config.getSparkStreamingKafkaMaxRatePerPartition()); // rate limiting
    sparkConf.setAppName("StreamingEngine-" + config.getTopicSet().toString() + "-" + config.getNamespace());

    if (config.getLocalMode()) {
        sparkConf.setMaster("local[4]");
    }
    return sparkConf;
}
 
开发者ID:ameyamk,项目名称:spark-streaming-direct-kafka,代码行数:12,代码来源:AbstractSparkLayer.java

示例8: run

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
@Override
public void run(String... args) throws Exception {
    SparkConf sparkConf = new SparkConf();
    sparkConf.set("spark.yarn.jar", config.getAssemblyJar());

    List<String> submitArgs = new ArrayList<String>();
    if (StringUtils.hasText(config.getAppName())) {
        submitArgs.add("--name");
        submitArgs.add(config.getAppName());
    }
    submitArgs.add("--jar");
    submitArgs.add(config.getAppJar());
    submitArgs.add("--class");
    submitArgs.add(config.getAppClass());
    if (StringUtils.hasText(config.getResourceFiles())) {
        submitArgs.add("--files");
        submitArgs.add(config.getResourceFiles());
    }
    if (StringUtils.hasText(config.getResourceArchives())) {
        submitArgs.add("--archives");
        submitArgs.add(config.getResourceArchives());
    }
    submitArgs.add("--executor-memory");
    submitArgs.add(config.getExecutorMemory());
    submitArgs.add("--num-executors");
    submitArgs.add("" + config.getNumExecutors());
    for (String arg : config.getAppArgs()) {
        submitArgs.add("--arg");
        submitArgs.add(arg);
    }
    logger.info("Submit App with args: " + Arrays.asList(submitArgs));
    ClientArguments clientArguments =
            new ClientArguments(submitArgs.toArray(new String[submitArgs.size()]), sparkConf);
    clientArguments.isClusterMode();
    Client client = new Client(clientArguments, hadoopConfiguration, sparkConf);
    System.setProperty("SPARK_YARN_MODE", "true");
    try {
        client.run();
    } catch (Throwable t) {
        logger.error("Spark Application failed: " + t.getMessage(), t);
        throw new RuntimeException("Spark Application failed", t);
    }
}
 
开发者ID:spring-cloud-task-app-starters,项目名称:spark-yarn,代码行数:44,代码来源:SparkYarnTaskConfiguration.java

示例9: setupTest

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
@After
@Before
public void setupTest() {
    SparkConf sparkConfiguration = new SparkConf();
    sparkConfiguration.setAppName(this.getClass().getCanonicalName() + "-setupTest");
    sparkConfiguration.set("spark.master", "local[4]");
    JavaSparkContext sparkContext = new JavaSparkContext(SparkContext.getOrCreate(sparkConfiguration));
    sparkContext.close();
    Spark.create(sparkContext.sc());
    Spark.close();
    logger.info("SparkContext has been closed for " + this.getClass().getCanonicalName() + "-setupTest");
}
 
开发者ID:PKUSilvester,项目名称:LiteGraph,代码行数:13,代码来源:AbstractSparkTest.java

示例10: createSparkContext

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
private static JavaSparkContext createSparkContext(SparkContextOptions contextOptions) {
  if (usesProvidedSparkContext) {
    LOG.info("Using a provided Spark Context");
    JavaSparkContext jsc = contextOptions.getProvidedSparkContext();
    if (jsc == null || jsc.sc().isStopped()){
      LOG.error("The provided Spark context " + jsc + " was not created or was stopped");
      throw new RuntimeException("The provided Spark context was not created or was stopped");
    }
    return jsc;
  } else {
    LOG.info("Creating a brand new Spark Context.");
    SparkConf conf = new SparkConf();
    if (!conf.contains("spark.master")) {
      // set master if not set.
      conf.setMaster(contextOptions.getSparkMaster());
    }

    if (contextOptions.getFilesToStage() != null && !contextOptions.getFilesToStage().isEmpty()) {
      conf.setJars(contextOptions.getFilesToStage().toArray(new String[0]));
    }

    conf.setAppName(contextOptions.getAppName());
    // register immutable collections serializers because the SDK uses them.
    conf.set("spark.kryo.registrator", BeamSparkRunnerRegistrator.class.getName());
    return new JavaSparkContext(conf);
  }
}
 
开发者ID:apache,项目名称:beam,代码行数:28,代码来源:SparkContextFactory.java

示例11: main

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
public static void main(String[] args) {

        SparkConf sparkConf = new SparkConf();//.setAppName("BLASpark - Example CG");

        sparkConf.set("spark.shuffle.reduceLocality.enabled","false");
        //sparkConf.set("spark.memory.useLegacyMode","true");


        //The ctx is created from the previous config
        JavaSparkContext ctx = new JavaSparkContext(sparkConf);
        //ctx.hadoopConfiguration().set("parquet.enable.summary-metadata", "false");

        GeneralOptions BLASparkOptions = new GeneralOptions(args);

        if(BLASparkOptions.getMode() == GeneralOptions.Mode.DMXV) {

            LOG.warn("Starting dense matrix dot vector multiplication example...");
            DmXV exDmXV = new DmXV(BLASparkOptions, ctx);

            exDmXV.calculate();
        }
        else if(BLASparkOptions.getMode() == GeneralOptions.Mode.CG) {

            LOG.warn("Starting conjugate gradient example...");
            ConjugateGradientExample exCG = new ConjugateGradientExample(BLASparkOptions, ctx);

            exCG.calculate();
        }
        else if(BLASparkOptions.getMode() == GeneralOptions.Mode.JACOBI) {

            LOG.warn("Starting Jacobi example...");
            JacobiExample exJacobi = new JacobiExample(BLASparkOptions, ctx);

            exJacobi.calculate();
        }
        else if(BLASparkOptions.getMode() == GeneralOptions.Mode.DMXDM) {

            LOG.warn("Starting dense matrix dot dense matrix example...");
            DmXDm exDmXDm = new DmXDm(BLASparkOptions, ctx);

            exDmXDm.calculate();
        }
        else {
            LOG.warn("No execution mode selected...");
            BLASparkOptions.printHelp();
        }

    }
 
开发者ID:jmabuin,项目名称:BLASpark,代码行数:49,代码来源:BLASpark.java

示例12: initialize

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
private LinkedBlockingQueue<Kryo> initialize(final Configuration conf) {
    // DCL is safe in this case due to volatility
    if (!initialized) {
        synchronized (UnshadedKryoShimService.class) {
            if (!initialized) {
                final SparkConf sparkConf = new SparkConf();

                // Copy the user's IoRegistry from the param conf to the SparkConf we just created
                final String regStr = conf.getString(GryoPool.CONFIG_IO_REGISTRY);
                if (null != regStr) { // SparkConf rejects null values with NPE, so this has to be checked before set(...)
                    sparkConf.set(GryoPool.CONFIG_IO_REGISTRY, regStr);
                }
                // Setting spark.serializer here almost certainly isn't necessary, but it doesn't hurt
                sparkConf.set("spark.serializer", IoRegistryAwareKryoSerializer.class.getCanonicalName());

                final String registrator = conf.getString("spark.kryo.registrator");
                if (null != registrator) {
                    sparkConf.set("spark.kryo.registrator", registrator);
                    log.info("Copied spark.kryo.registrator: {}", registrator);
                } else {
                    log.info("Not copying spark.kryo.registrator");
                }

                // Reuse Gryo poolsize for Kryo poolsize (no need to copy this to SparkConf)
                final int poolSize = conf.getInt(GryoPool.CONFIG_IO_GRYO_POOL_SIZE,
                        GryoPool.CONFIG_IO_GRYO_POOL_SIZE_DEFAULT);
                // Instantiate the spark.serializer
                final IoRegistryAwareKryoSerializer ioReg = new IoRegistryAwareKryoSerializer(sparkConf);
                // Setup a pool backed by our spark.serializer instance

                for (int i = 0; i < poolSize; i++) {
                    KRYOS.add(ioReg.newKryo());
                }

                initialized = true;
            }
        }
    }

    return KRYOS;
}
 
开发者ID:PKUSilvester,项目名称:LiteGraph,代码行数:42,代码来源:UnshadedKryoShimService.java

示例13: ComputeResponse

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
public ComputeResponse(FileSystem fileSys) throws PIRException
{
  fs = fileSys;
  storage = new HadoopFileSystemStore(fs);

  dataInputFormat = SystemConfiguration.getProperty("pir.dataInputFormat");
  if (!InputFormatConst.ALLOWED_FORMATS.contains(dataInputFormat))
  {
    throw new IllegalArgumentException("inputFormat = " + dataInputFormat + " is of an unknown form");
  }
  logger.info("inputFormat = " + dataInputFormat);
  if (dataInputFormat.equals(InputFormatConst.BASE_FORMAT))
  {
    inputData = SystemConfiguration.getProperty("pir.inputData", "none");
    if (inputData.equals("none"))
    {
      throw new IllegalArgumentException("For inputFormat = " + dataInputFormat + " an inputFile must be specified");
    }
    logger.info("inputFile = " + inputData);
  }
  else if (dataInputFormat.equals(InputFormatConst.ES))
  {
    esQuery = SystemConfiguration.getProperty("pir.esQuery", "none");
    esResource = SystemConfiguration.getProperty("pir.esResource", "none");
    if (esQuery.equals("none"))
    {
      throw new IllegalArgumentException("esQuery must be specified");
    }
    if (esResource.equals("none"))
    {
      throw new IllegalArgumentException("esResource must be specified");
    }
    logger.info("esQuery = " + esQuery + " esResource = " + esResource);
  }
  outputFile = SystemConfiguration.getProperty("pir.outputFile");
  outputDirExp = outputFile + "_exp";

  queryInput = SystemConfiguration.getProperty("pir.queryInput");
  String stopListFile = SystemConfiguration.getProperty("pir.stopListFile");
  useModExpJoin = SystemConfiguration.getBooleanProperty("pir.useModExpJoin", false);

  logger.info("outputFile = " + outputFile + " queryInputDir = " + queryInput + " stopListFile = " + stopListFile + " esQuery = " + esQuery + " esResource = "
      + esResource);

  // Set the necessary configurations
  SparkConf conf = new SparkConf().setAppName("SparkPIR").setMaster("yarn-cluster");
  conf.set("es.nodes", SystemConfiguration.getProperty("es.nodes", "none"));
  conf.set("es.port", SystemConfiguration.getProperty("es.port", "none"));
  conf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer");
  conf.set("spark.memory.storageFraction", "0.10");
  conf.set("spark.memory.fraction", "0.25");
  // conf.set("spark.memory.fraction", "0.25");
  // conf.set("spark.executor.extraJavaOptions", "-XX:+UseCompressedOops");
  sc = new JavaSparkContext(conf);

  // Setup, run query, teardown
  logger.info("Setting up for query run");
  try
  {
    setup();
  } catch (IOException e)
  {
    throw new PIRException("An error occurred setting up the Spark responder.", e);
  }
  logger.info("Setup complete");
}
 
开发者ID:apache,项目名称:incubator-pirk,代码行数:67,代码来源:ComputeResponse.java

示例14: ComputeStreamingResponse

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
public ComputeStreamingResponse(FileSystem fileSys) throws PIRException
{
  fs = fileSys;
  storage = new HadoopFileSystemStore(fs);

  dataInputFormat = SystemConfiguration.getProperty("pir.dataInputFormat");
  if (!InputFormatConst.ALLOWED_FORMATS.contains(dataInputFormat))
  {
    throw new IllegalArgumentException("inputFormat = " + dataInputFormat + " is of an unknown form");
  }
  logger.info("inputFormat = " + dataInputFormat);
  if (dataInputFormat.equals(InputFormatConst.BASE_FORMAT))
  {
    inputData = SystemConfiguration.getProperty("pir.inputData", "none");
    if (inputData.equals("none"))
    {
      throw new IllegalArgumentException("For inputFormat = " + dataInputFormat + " an inputFile must be specified");
    }
    logger.info("inputFile = " + inputData);
  }
  else if (dataInputFormat.equals(InputFormatConst.ES))
  {
    esQuery = SystemConfiguration.getProperty("pir.esQuery", "none");
    esResource = SystemConfiguration.getProperty("pir.esResource", "none");
    if (esQuery.equals("none"))
    {
      throw new IllegalArgumentException("esQuery must be specified");
    }
    if (esResource.equals("none"))
    {
      throw new IllegalArgumentException("esResource must be specified");
    }
    logger.info("esQuery = " + esQuery + " esResource = " + esResource);
  }
  outputFile = SystemConfiguration.getProperty("pir.outputFile");
  outputDirExp = outputFile + "_exp";

  queryInput = SystemConfiguration.getProperty("pir.queryInput");
  String stopListFile = SystemConfiguration.getProperty("pir.stopListFile");

  logger.info("outputFile = " + outputFile + " queryInputDir = " + queryInput + " stopListFile = " + stopListFile + " esQuery = " + esQuery + " esResource = "
      + esResource);

  // Pull the batchSeconds and windowLength parameters
  long batchSeconds = SystemConfiguration.getLongProperty("pir.sparkstreaming.batchSeconds", 30);
  windowLength = SystemConfiguration.getLongProperty("pir.sparkstreaming.windowLength", 60);
  if (windowLength % batchSeconds != 0)
  {
    throw new IllegalArgumentException("batchSeconds = " + batchSeconds + " must divide windowLength = " + windowLength);
  }
  useQueueStream = SystemConfiguration.getBooleanProperty("pir.sparkstreaming.useQueueStream", false);
  logger.info("useQueueStream = " + useQueueStream);

  // Set the necessary configurations
  SparkConf conf = new SparkConf().setAppName("SparkPIR").setMaster("yarn-cluster");
  conf.set("es.nodes", SystemConfiguration.getProperty("es.nodes", "none"));
  conf.set("es.port", SystemConfiguration.getProperty("es.port", "none"));
  conf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer");
  conf.set("spark.streaming.stopGracefullyOnShutdown", SystemConfiguration.getProperty("spark.streaming.stopGracefullyOnShutdown", "false"));

  JavaSparkContext sc = new JavaSparkContext(conf);
  jssc = new JavaStreamingContext(sc, Durations.seconds(batchSeconds));

  // Setup, run query, teardown
  logger.info("Setting up for query run");
  try
  {
    setup();
  } catch (IOException e)
  {
    throw new PIRException("An error occurred setting up the streaming responder.", e);
  }
  logger.info("Setup complete");
}
 
开发者ID:apache,项目名称:incubator-pirk,代码行数:75,代码来源:ComputeStreamingResponse.java

示例15: main

import org.apache.spark.SparkConf; //导入方法依赖的package包/类
@SuppressWarnings("deprecation")
public static void main(String[] args) {
	System.setProperty("hadoop.home.dir", "C:\\softwares\\Winutils");
	SparkConf conf =new SparkConf().setMaster("local").setAppName("Cassandra Example");
	conf.set("spark.cassandra.connection.host", "127.0.0.1");
	//conf.set("spark.sql.warehouse.dir", "file:////C:/Users/sgulati/spark-warehouse");
	
	JavaSparkContext jsc=new JavaSparkContext(conf);
	JavaRDD<Employee> cassandraTable = CassandraJavaUtil.javaFunctions(jsc).cassandraTable("my_keyspace", "emp",CassandraJavaUtil.mapRowTo(Employee.class));
	
    JavaRDD<String> selectEmpDept = CassandraJavaUtil.javaFunctions(jsc).cassandraTable("my_keyspace", "emp",CassandraJavaUtil.mapColumnTo(String.class)).select("emp_dept","emp_name");
	
	 
    cassandraTable.collect().forEach(System.out::println);
    //selectEmpDept.collect().forEach(System.out::println);
	
    
    CassandraJavaUtil.javaFunctions(cassandraTable)
       .writerBuilder("my_keyspace", "emp1", CassandraJavaUtil.mapToRow(Employee.class)).saveToCassandra();
    
	/*SQLContext sqlContext = new SQLContext(jsc);
	
	Map<String,String> map =new HashMap<>();
	map.put("table" , "emp");
	map.put("keyspace", "my_keyspace");
	
	Dataset<Row> df = sqlContext.read().format("org.apache.spark.sql.cassandra")
	  .options(map)
	  .load();
	
	df.show();*/
	
	
	
}
 
开发者ID:PacktPublishing,项目名称:Apache-Spark-2x-for-Java-Developers,代码行数:36,代码来源:CassandraExample.java


注:本文中的org.apache.spark.SparkConf.set方法示例由纯净天空整理自Github/MSDocs等开源代码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。