Skip to content

SparkSQL Implementation#1146

Merged
rfecher merged 6 commits intomasterfrom
sparksql
Aug 7, 2017
Merged

SparkSQL Implementation#1146
rfecher merged 6 commits intomasterfrom
sparksql

Conversation

@blastarr
Copy link
Copy Markdown
Contributor

@blastarr blastarr commented Aug 2, 2017

No description provided.

# Conflicts:
#	analytics/javaspark/src/main/java/mil/nga/giat/geowave/analytic/javaspark/GeoWaveRDD.java
#	analytics/javaspark/src/main/resources/META-INF/services/mil.nga.giat.geowave.core.cli.spi.CLIOperationProviderSpi
#	extensions/adapters/vector/src/main/java/mil/nga/giat/geowave/adapter/vector/util/FeatureDataUtils.java
#	extensions/adapters/vector/src/main/java/mil/nga/giat/geowave/adapter/vector/utils/DateUtilities.java
#	extensions/adapters/vector/src/main/java/mil/nga/giat/geowave/adapter/vector/utils/GeometryUtils.java
#	pom.xml
#	test/src/test/java/mil/nga/giat/geowave/test/javaspark/GeoWaveJavaSparkIT.java
#	test/src/test/java/mil/nga/giat/geowave/test/javaspark/GeoWaveJavaSparkKMeansIT.java
import javax.measure.unit.Unit;

import mil.nga.giat.geowave.adapter.vector.plugin.GeoWaveGTDataStore;
import mil.nga.giat.geowave.adapter.vector.stats.FeatureBoundingBoxStatistics;
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

imports not used

}

public static String getFirstTimeField(
final DataStorePluginOptions dataStore,
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

not used


public static ByteArrayId getFirstFeatureAdapter(
DataStorePluginOptions dataStore ) {
CloseableIterator<DataAdapter<?>> adapterIt = dataStore.createAdapterStore().getAdapters();
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

non-deterministic behavior

private final static Logger LOGGER = LoggerFactory.getLogger(GeoWaveJavaSparkSQLIT.class);

@GeoWaveTestStore(value = {
// GeoWaveStoreType.ACCUMULO,
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

why no accumulo?

@@ -0,0 +1,3 @@
{"name":"Michael"}
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

delete

@@ -0,0 +1,6 @@
0.0 0.0 0.0
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

delete

@@ -0,0 +1,368 @@
/*
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

delete

package mil.nga.giat.geowave.analytic.javaspark.kmeans;

import org.apache.spark.SparkConf;
import org.apache.spark.api.java.JavaSparkContext;
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

delete

@@ -0,0 +1,314 @@
package mil.nga.giat.geowave.analytic.javaspark;

Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

delete

UDF2<String, String, Boolean>
{
private final static Logger LOGGER = LoggerFactory.getLogger(GeomFunction.class);
protected GeomReader geomReader = new GeomReader();
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

this is why the reader must be serializable (I think marking this transient shouldn't require it to be serializable)

private static Logger LOGGER = LoggerFactory.getLogger(SimpleFeatureDataFrame.class);

private final StructType schema;
private final GeomWriter geomWriter = new GeomWriter();
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

again, this is where its required to be serializable (we can consider marking it transient)

@rfecher rfecher merged commit bcd8b4d into master Aug 7, 2017
@rfecher rfecher deleted the sparksql branch August 7, 2017 15:30
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants