NoSQL Databases. Big Data 2015
|
|
|
- Kellie Clarke
- 10 years ago
- Views:
Transcription
1 NoSQL Databases Big Data 2015
2 What s the matter
3 What s the matter
4 What s the matter
5 CAP Theorem
6 NoSQL Family
7 Which one?
8 NoSQL Family
9 Amazon DynamoDB
10 Amazon DynamoDB
11 Amazon DynamoDB Relational Data Model Concepts In a relational database, a table has a predefined schema the table name primary key list of its column names and their data types. All records stored in the table must have the same set of columns.
12 Amazon DynamoDB Bycicle id price title type brand color description Bike-205 Hybrid C {black, red D1 RDB Bike-203 Road B {black, green, red D Bike-202 Road A {black, green D Bike-201 Road D {black, red D Bike-204 Mountain B {red D5 Book id price title authors dimensions ISBN InPublicatio pages T1 {A1, A2 8.5 x 11.0 x n T2 {A1, A2, A3 8.5 x 11.0 x T3 {A1 8.5 x 11.0 x
13 Amazon DynamoDB Products id price category title type brand color descript authors dimensi ISBN InPublic pages Bycicle 20- Hybrid C {black, ion D1 NULL NULL ons NULL NULL ation NULL Bycicle Bike Road B {black, red D2 NULL NULL NULL NULL NULL Bycicle Bike Road A green, {black, red D3 NULL NULL NULL NULL NULL Bycicle Bike Road D {black, green D4 NULL NULL NULL NULL NULL Bycicle Bike Mountai B {red D5 NULL NULL NULL NULL NULL Book Bike-204 T1 NULL n NULL NULL NULL {A1, 8.5 x Book T2 NULL NULL NULL NULL {A1, A2 A2, x Book T3 NULL NULL NULL NULL {A1 A x 11.0 x RDB
14 Amazon DynamoDB Data Model Concepts a database is a collection of tables; a table is a collection of items; each item is a collection of attributes. A DynamoDB table is schema-less Individual items in a DynamoDB table can have any number of attributes Each attribute in an item is a name-value pair. An attribute can be single-valued or multi-valued set. There is a primary-key attribute.
15 Amazon DynamoDB For example, consider storing a catalog of products in DynamoDB.You can create a table, ProductCatalog, with the Id attribute as its primary key. ProductCatalog ( Id,... )
16 Amazon DynamoDB You can store various kinds of product items in the table. { " Id = 201" ProductName = "18-Bicycle 201"" Description = "201 description"" BicycleType = "Road"" Brand = "Brand-Company A"" Price = 100" Gender = "M"" Color = [ "Red", "Black" ]" ProductCategory = "Bicycle"" { Id = 101 ProductName = "Book 101 Title"" ISBN = " "" Authors = [ "Author 1", "Author 2" ]" Price = 2" Dimensions = "8.5 x 11.0 x 0.5"" PageCount = 500" InPublication = 1" ProductCategory = "Book""
17 Running Java Examples for DynamoDB Download and install the AWS Toolkit for Eclipse. This toolkit includes the AWS SDK for Java, along with preconfigured templates for building applications. From the Eclipse menu, create a new AWS Java Project You will now need to create a default credential profiles file. AwsCredentials.properties aws_access_key_id = <Your Access Key ID>" aws_secret_access_key = <Your Secret Key>
18 Running Java Examples for DynamoDB
19 Running Java Examples for DynamoDB static AmazonDynamoDBClient client; private static void init() throws Exception { /* * The ProfileCredentialsProvider will return your [default] * credential profile by reading from the credentials file located at * (~/.aws/credentials). */ AWSCredentials credentials = null; try { //credentials = new ProfileCredentialsProvider().getCredentials(); credentials = new PropertiesCredentials( AmazonDynamoDBSample.class.getResourceAsStream("AwsCredentials.properties")); catch (Exception e) { throw new AmazonClientException( "Cannot load the credentials from the credential profiles file. " + "Please make sure that your credentials file is at the correct " + "location (~/.aws/credentials), and is in valid format.", e); client = new AmazonDynamoDBClient(credentials); Region uswest2 = Region.getRegion(Regions.US_WEST_2); client.setregion(uswest2);
20 Running Java Examples for DynamoDB public class CreateTablesUploadData { static AmazonDynamoDBClient client; static String productcatalogtablename = ProductCatalog ;... public static void main(string[] args) throws Exception { try { init(); deletetable(productcatalogtablename); waitfortabletobedeleted(productcatalogtablename); // Parameter1: table name // Parameter2: reads per second // Parameter3: writes per second // Parameter4/5: hash key and type // Parameter6/7: range key and type (if applicable) createtable(productcatalogtablename, 10L, 5L, "Id", N"); waitfortabletobecomeavailable(productcatalogtablename); uploadsampleproducts(productcatalogtablename); catch (AmazonServiceException ase) { System.err.println("Data load script failed: " + ase); ase.printstacktrace();
21 Running Java Examples for DynamoDB private static void createtable(string tablename, long readcapacityunits, long writecapacityunits, String hashkeyname, String hashkeytype) { createtable(tablename, readcapacityunits, writecapacityunits, hashkeyname, hashkeytype, null, null); private static void createtable(string tablename, long readcapacityunits, long writecapacityunits, String hashkeyname, String hashkeytype, String rangekeyname, String rangekeytype) { try { System.out.println("Creating table " + tablename); ArrayList<KeySchemaElement> ks = new ArrayList<KeySchemaElement>(); ArrayList<AttributeDefinition> attributedefinitions = new ArrayList<AttributeDefinition>(); ks.add(new KeySchemaElement().withAttributeName(hashKeyName).withKeyType(KeyType.HASH)); attributedefinitions.add(new AttributeDefinition().withAttributeName( hashkeyname).withattributetype(hashkeytype)); if (rangekeyname = null){ ks.add(new KeySchemaElement().withAttributeName( rangekeyname).withkeytype(keytype.range)); attributedefinitions.add(new AttributeDefinition().withAttributeName( rangekeyname).withattributetype(rangekeytype));
22 Running Java Examples for DynamoDB CreateTableRequest request = new CreateTableRequest().withTableName(tableName).withKeySchema(ks).withProvisionedThroughput(provisionedthroughput); request.setattributedefinitions(attributedefinitions); client.createtable(request); catch (AmazonServiceException ase) { System.err.println("Failed to create table " + tablename + " " + ase);
23 Running Java Examples for DynamoDB private static void uploadsampleproducts(string tablename) { try { // Add books. Map<String, AttributeValue> item = new HashMap<String, AttributeValue>(); item.put("id", new AttributeValue().withN("101")); item.put("title", new AttributeValue().withS("Book 101 Title")); item.put("isbn", new AttributeValue().withS(" ")); item.put("authors", new AttributeValue().withSS(Arrays.asList("Author1"))); item.put("price", new AttributeValue().withN("2")); item.put("dimensions", new AttributeValue().withS("8.5 x 11.0 x 0.5")); item.put("pagecount", new AttributeValue().withN("500")); item.put("inpublication", new AttributeValue().withN("1")); item.put("productcategory", new AttributeValue().withS("Book")); PutItemRequest itemrequest = new PutItemRequest().withTableName(tableName).withItem(item); client.putitem(itemrequest); item.clear(); // Add Bicycle. item.put("id", new AttributeValue().withN("201")); item.put("title", new AttributeValue().withS("18-Bike-201")); // Size, followed by some title. item.put("description", new AttributeValue().withS("201 Description")); item.put("bicycletype", new AttributeValue().withS("Road")); item.put("brand", new AttributeValue().withS("Mountain A")); // Trek, Specialized. item.put("price", new AttributeValue().withN("100")); item.put("gender", new AttributeValue().withS("M")); // Men's item.put("color", new AttributeValue().withSS(Arrays.asList("Red", "Black"))); item.put("productcategory", new AttributeValue().withS( Bicycle")); itemrequest = new PutItemRequest().withTableName(tableName).withItem(item); client.putitem(itemrequest); catch (AmazonServiceException ase) { System.err.println("Failed to create item in " + tablename);
24 Running Java Examples for DynamoDB public class ObjectPersistenceQueryScanExample { static AmazonDynamoDBClient client;... public static void main(string[] args) throws Exception { try { init(); DynamoDBMapper mapper = new DynamoDBMapper(client); // Get a book - Id=101 GetBook(mapper, 101); catch (Throwable t) { System.err.println("Error running the ObjectPersistenceQueryScanExample: " + t); t.printstacktrace(); private static void GetBook(DynamoDBMapper mapper, int id) throws Exception { System.out.println("GetBook: Get book Id='101' "); System.out.println("Book table has no range key attribute, so you Get (but no query)."); Book book = mapper.load(book.class, 101); System.out.format("Id = %s Title = %s, ISBN = %s %n", book.getid(), book.gettitle(), book.getisbn() );
25 Running Java Examples for public static class Book { private int id; private String title; private String ISBN; private int price; private int pagecount; private String productcategory; private int public int getid() { return id; public void setid(int id) { this.id = public String gettitle() { return title; public void settitle(string title) { this.title = public String getisbn() { return ISBN; public void setisbn(string ISBN) { this.isbn = public int getprice() { return price; public void setprice(int price) { this.price = public int getpagecount() { return pagecount; public void setpagecount(int pagecount) { this.pagecount = public String getproductcategory() { return productcategory; public void setproductcategory(string productcategory) { this.productcategory = public int getinpublication() { return inpublication; public void setinpublication(int inpublication) { this.inpublication = inpublication;
26 HBase Configuration In the bash_profile export all needed environment variables
27 HBase Configuration Download a stable release of apache HBase: hbase tar.gz
28 HBase Configuration Go to the conf directory in the hbase-home directory set the hbase-env.sh file as follows
29 HBase Configuration Go to the conf directory in the hbase-home directory set the hbase-site.xml file as follows
30 HBASE Running Running hadoop: Running HBase: $:~hbase-*/bin/start-hbase.sh Hbase shell: $:~hbase-*/bin/hbase shell HBase Shell; enter 'help<return>' for list of supported commands. Type "exit<return>" to leave the HBase Shell Version , r , Tue Apr 22 00:21:01 UTC 2014 hbase(main):001:0>
31 Apache HBase HBase is built on top of HDFS HBase files are internally stored in HDFS
32 Apache HBase
33 Apache HBase HBase is based on Google s Bigtable model (Key-Value pairs) Column Family Row key TimeStamp value
34 Logical View of HBase Apache HBase
35 Apache HBase Keys and Column Families Each row has a Key Each record is divided into Column Families Each column family consists of one or more Columns
36 Tables can be very sparse Apache HBase
37 Apache HBase The example of products { " Id = 201" ProductName = "18-Bicycle 201"" Description = "201 description"" BicycleType = "Road"" Brand = "Brand-Company A"" Price = 100" Color = [ "Red", "Black" ]" ProductCategory = "Bicycle"" { Id = 101 ProductName = "Book 101 Title"" ISBN = " "" Authors = [ "Author 1", "Author 2" ]" Price = 2" Dimensions = "8.5 x 11.0 x 0.5"" PageCount = 500" InPublication = 1" ProductCategory = "Book""
38 Apache HBase Products row key t general bike book row1 t1 general:id = 201" general:productname " = "18-Bicycle 201" bike:description = "D1"" bike:bicycletype = "Road"" bike:brand = "Brand-Company A"" bike:price = 100" bike:color = [ "Red", "Black" ] row2 t2 general:id = 101" general:productname " = Book 101 Title" book:isbn = " "" book:authors = ["A1", "A2"]" book:price = 2" book:dimensions = 8.5 x 11.0 x 0.5"" book:pages = 500" book:inpublication = 1
39 HBASE Running Hbase shell: hbase(main):005:0> create 'products', 'general', 'bike', 'book' 0 row(s) in seconds hbase(main):006:0> list TABLE products 1 row(s) in seconds hbase(main):007:0> put 'products', 'row1', 'general:id', row(s) in seconds hbase(main):009:0> put 'products', 'row1', 'general:productname', '18-Bicycle 201' 0 row(s) in seconds hbase(main):011:0> scan 'products' ROW COLUMN+CELL row1 column=general:id, timestamp= , value=201 row1 column=general:productname, timestamp= , value=18-bicycle row(s) in seconds
40 HBASE Running Hbase shell: hbase(main):007:0> put 'products', 'row2', 'general:id', row(s) in seconds hbase(main):009:0> put 'products', 'row2', 'general:productname', Book 101 Title' 0 row(s) in seconds hbase(main):014:0> scan 'products' ROW COLUMN+CELL row1 column=general:id, timestamp= , value=201 row1 column=general:productname, timestamp= , value=18-bicycle 201 row2 column=general:id, timestamp= , value=101 row2 column=general:productname, timestamp= , value=book 101 Title 2 row(s) in seconds
41 HBASE Running Hbase shell: hbase(main):015:0> get 'products', 'row1' COLUMN CELL general:id timestamp= , value=201 general:productname timestamp= , value=18-bicycle row(s) in seconds hbase(main):016:0> exit Stopping HBase: $:~hbase-*/bin/stop-hbase.sh stopping hbase...
42 HBASE in Java Hbase Java Project: create a lib folder and copy that jars from the home-folder of hbase
43 HBASE in Java Hbase Java Project: copy the hbase-site.xml file from the home-folder of hbase <configuration> <property> <name>hbase.rootdir</name> <value>hdfs://localhost:9000/hbase</value> </property> </configuration>
44 HBASE in Java HBaseTest.java: public class HBaseTest { private static Configuration conf = null; /** * Initialization */ static { conf = HBaseConfiguration.create(); public static void main(string[] agrs) { try { String tablename = "scores"; String[] familys = { "grade", "course" ; HBaseTest.creatTable(tablename, familys); // add record zkb HBaseTest.addRecord(tablename, "zkb", "grade", "", "5"); HBaseTest.addRecord(tablename, "zkb", "course", "", "90"); HBaseTest.addRecord(tablename, "zkb", "course", "math", "97"); HBaseTest.addRecord(tablename, "zkb", "course", "art", "87"); // add record baoniu HBaseTest.addRecord(tablename, "baoniu", "grade", "", "4"); HBaseTest.addRecord(tablename, "baoniu", "course", "math", "89"); start HBase.... catch (Exception e) { e.printstacktrace();
45 HBASE in Java HBaseTest.java: public static void creattable(string tablename, String[] familys) throws Exception { HBaseAdmin admin = new HBaseAdmin(conf); if (admin.tableexists(tablename)) { System.out.println("table already exists"); else { HTableDescriptor tabledesc = new HTableDescriptor(tableName); for (int i = 0; i < familys.length; i++) { tabledesc.addfamily(new HColumnDescriptor(familys[i])); admin.createtable(tabledesc); System.out.println("create table " + tablename + " ok."); public static void addrecord(string tablename, String rowkey, String family, String qualifier, String value) throws Exception { try { HTable table = new HTable(conf, tablename); Put put = new Put(Bytes.toBytes(rowKey)); put.add(bytes.tobytes(family), Bytes.toBytes(qualifier), Bytes.toBytes(value)); table.put(put); System.out.println("insert recored " + rowkey + " to table + tablename + " ok."); catch (IOException e) { e.printstacktrace();
46 HBASE in Java HBaseTest.java: public class HBaseTest { private static Configuration conf = null; /** * Initialization */ static { conf = HBaseConfiguration.create(); public static void main(string[] agrs) { try { start HBase.... System.out.println("===========get one record========"); HBaseTest.getOneRecord(tablename, "zkb"); System.out.println("===========show all record========"); HBaseTest.getAllRecord(tablename); System.out.println("===========del one record========"); HBaseTest.delRecord(tablename, "baoniu"); HBaseTest.getAllRecord(tablename); System.out.println("===========show all record========"); HBaseTest.getAllRecord(tablename); catch (Exception e) { e.printstacktrace();
47 HBASE in Java HBaseTest.java: public static void getonerecord (String tablename, String rowkey) throws IOException{ HTable table = new HTable(conf, tablename); Get get = new Get(rowKey.getBytes()); Result rs = table.get(get); for(keyvalue kv : rs.raw()){ System.out.print(new String(kv.getRow()) + " " ); System.out.print(new String(kv.getFamily()) + ":" ); System.out.print(new String(kv.getQualifier()) + " " ); System.out.print(kv.getTimestamp() + " " ); System.out.println(new String(kv.getValue())); public static void delrecord(string tablename, String rowkey) throws IOException { HTable table = new HTable(conf, tablename); List<Delete> list = new ArrayList<Delete>(); Delete del = new Delete(rowKey.getBytes()); list.add(del); table.delete(list); System.out.println("del recored " + rowkey + " ok.");
48 Cassandra Configuration Download a release of apache Cassandra: apache-cassandra bin.tar.gz
49 Cassandra Configuration Go to the conf directory in the cassandra-home directory set the cassandra.yaml file as follows # Directories where Cassandra should store data on disk. Cassandra # will spread data evenly across them, subject to the granularity of # the configured compaction strategy. data_file_directories: <home>/dsc-cassandra-2.0.7/lib/cassandra/data # commit log commitlog_directory: <home>/dsc-cassandra-2.0.7/lib/cassandra/commitlog # saved caches saved_caches_directory: <home>/dsc-cassandra-2.0.7/lib/cassandra/saved_caches
50 Cassandra Configuration Go to the conf directory in the cassandra-home directory set the log4j-server.properties file as follows log4j.appender.r.file=<home>/dsc-cassandra-2.0.7/log/cassandra/system.log
51 Cassandra Configuration Go to the cassandra-home directory create the following directories
52 Cassandra Running Running cassandra: $:~cassandra-*/bin/cassandra -f Warning: The service should start in the foreground and log gratuitously to the console. Assuming you don't see messages with scary words like "error", or "fatal", or anything that looks like a Java stack trace, then everything should be working. You have to keep open the terminal window and execute cassandra commands in another terminal window. Press "Control-C" in the first terminal window to stop Cassandra.
53 Cassandra Running Running cassandra: $:~cassandra-*/bin/cassandra Warning: If you start up Cassandra without the "-f" option, it will run in the background. You can stop the process by killing it, using 'pkill -f CassandraDaemon'
54 Cassandra Running Running cqlsh: $:~cassandra-*/bin/cqlsh it is an interactive command line interface for Cassandra. cqlsh allows you to execute CQL (Cassandra Query Language) statements against Cassandra. Using CQL, you can define a schema, insert data, execute queries. Connected to Test Cluster at localhost:9160. [cqlsh Cassandra CQL spec Thrift protocol ] Use HELP for help. cqlsh>
55 Apache Cassandra Cassandra is quite similar to HBase keyspace( column(family( se)ngs( se)ngs( column( name( value( 3mestamp(
56 Apache Cassandra The example of products { " Id = 201" ProductName = "18-Bicycle 201"" Description = "201 description"" BicycleType = "Road"" Brand = "Brand-Company A"" Price = 100" Color = "Red"" ProductCategory = "Bicycle"" { Id = 101 ProductName = "Book 101 Title"" ISBN = " "" Authors = "Author 1"" Price = 2" Dimensions = "8.5 x 11.0 x 0.5"" PageCount = 500" InPublication = 1" ProductCategory = "Book""
57 Apache Cassandra keyspace: Products columnfamily: Bycicle id price title type brand color description Bike-201 Road D {black, red D4 columnfamily: Book id price title authors dimensions ISBN InPublication pages T3 {A1 8.5 x 11.0 x
58 Cassandra Running Connected to Test Cluster at localhost:9160. [cqlsh Cassandra CQL spec Thrift protocol ] Use HELP for help. cqlsh> CREATE KEYSPACE products WITH REPLICATION = { 'class' : 'SimpleStrategy', 'replication_factor' : 1 ; cqlsh> USE products; cqlsh:products> CREATE TABLE bycicle ( id int PRIMARY KEY, price int, title text, type text, brand text, color text, description text ); cqlsh:products> CREATE COLUMNFAMILY bycicle ( id int PRIMARY KEY, price int, title text, type text, brand text, color text, description text );
59 Cassandra Running cqlsh:products> INSERT INTO bycicle (id, price, title, type, brand, color, description) VALUES (201, 100, '18-Bike-201', 'Road', 'D', 'red', D4'); cqlsh:products> SELECT * FROM bycicle; id brand color description price title type D red D Bike-201 Road (1 rows) cqlsh:products> ALTER COLUMNFAMILY bycicle ADD nation text; cqlsh:products> SELECT * FROM bycicle; id brand color description nation price title type D red D4 null Bike-201 Road (1 rows)
60 Cassandra in Java Cassandra Maven Java Project: create a Maven Java Project <project...> <modelversion>4.0.0</modelversion> <groupid>com.example.cassandra</groupid> <artifactid>your-project-name</artifactid> <version>0.0.1-snapshot</version> <dependencies> <dependency> <groupid>com.datastax.cassandra</groupid> <artifactid>cassandra-driver-core</artifactid> <version>2.0.2</version> </dependency> </dependencies> </project>
61 Cassandra in Java Cassandra Maven Java Project: create a Maven Java Project SELECT_ALL=SELECT * FROM dmloperations; INSERT_ALL = insert into dmloperations (dept_name, emp_id, emp_name ) VALUES (?,?,?); UPDATE_NAME = update dmloperations SET emp_name =? where dept_name=? and emp_id=?; DELETE_EMP = delete from dmloperations where dept_name=? and emp_id=?; SERVER_IP= keyspace=dmlkeyspace
62 Cassandra in Java DMLOperations.java public class DMLOperations { private static Properties properties; Cluster cluster = null; Session session = null; /** * Disconnect from the current cluster */ public void disconnect() { this.session.close(); this.cluster.close(); System.out.println("DisConnected"); /** ip keyspace * Connected to the keyspace and node */ public void connect(string ip, String keyspace) { this.cluster = Cluster.builder().addContactPoints(ip).build(); this.session = cluster.connect(keyspace); System.out.println("Connected"); start Cassandra
63 Cassandra in Java DMLOperations.java public static void main(string[] args) { DMLOperations object = new DMLOperations(); properties = object.loadproperties("data/queries"); object.connect(properties.getproperty("server_ip"), properties.getproperty("keyspace")); object.insertall("bigdata", 03, "sam"); object.insertall("bigdata", 05, "santhosh"); object.insertall("java", 04, "joe"); System.err.println("Inserted "); object.selectall(); object.update("bigdata", "samkrish", 03); System.err.println("Updated "); object.selectall(); object.delete("bigdata", 03); System.err.println("Deleted"); object.selectall(); object.disconnect();
64 Cassandra in Java DMLOperations.java /** propertiesfilename java.util.properties Object Load the values from File to Property * Object */ private Properties loadproperties(string propertiesfilename) { Properties prop = new Properties(); try { prop.load(new FileInputStream(propertiesFileName + ".properties")); catch (IOException ex) { ex.printstacktrace(); System.err.println(ex.getMessage()); return prop;
65 Cassandra in Java DMLOperations.java public static void main(string[] args) { DMLOperations object = new DMLOperations(); properties = object.loadproperties("data/queries"); object.connect(properties.getproperty("server_ip"), properties.getproperty("keyspace")); object.insertall("bigdata", 03, "sam"); object.insertall("bigdata", 05, "santhosh"); object.insertall("java", 04, "joe"); System.err.println("Inserted "); object.selectall(); object.update("bigdata", "samkrish", 03); System.err.println("Updated "); object.selectall(); object.delete("bigdata", 03); System.err.println("Deleted"); object.selectall(); object.disconnect(); Assume in Cassandra a keyspace dmlkeyspace and a table dmloperations ( dept_name, emp_id, emp_name )
66 Cassandra in Java /** DMLOperations.java INSERT_ALL = insert into dmloperations (dept_name, emp_id, emp_name ) VALUES (?,?,?); deptname EmployeeId * insert the data to column family */ public void insertall(string deptname, int EmployeeId, String EmployeeName) { BoundStatement boundstatement = null; PreparedStatement prepare_statement = null; prepare_statement = this.session.prepare(properties.getproperty("insert_all")); boundstatement = new BoundStatement(prepare_statement); this.session.execute(boundstatement.bind(deptname, EmployeeId, EmployeeName));
67 Cassandra in Java DMLOperations.java public static void main(string[] args) { DMLOperations object = new DMLOperations(); properties = object.loadproperties("data/queries"); object.connect(properties.getproperty("server_ip"), properties.getproperty("keyspace")); object.insertall("bigdata", 03, "sam"); object.insertall("bigdata", 05, "santhosh"); object.insertall("java", 04, "joe"); System.err.println("Inserted "); object.selectall(); object.update("bigdata", "samkrish", 03); System.err.println("Updated "); object.selectall(); object.delete("bigdata", 03); System.err.println("Deleted"); object.selectall(); object.disconnect();
68 Cassandra in Java /** DMLOperations.java SELECT_ALL = SELECT * FROM dmloperations; * Select all the rows from the given columnfamily */ public void selectall() { BoundStatement boundstatement = null; PreparedStatement prepare_statement = null; prepare_statement = this.session.prepare(properties.getproperty("select_all")); boundstatement = new BoundStatement(prepare_statement); ResultSet rs = this.session.execute(boundstatement); for (Row row : rs) { System.out.println(row.toString());
69 Resources
70 Resources
71 NoSQL Databases Big Data 2015
Xiaoming Gao Hui Li Thilina Gunarathne
Xiaoming Gao Hui Li Thilina Gunarathne Outline HBase and Bigtable Storage HBase Use Cases HBase vs RDBMS Hands-on: Load CSV file to Hbase table with MapReduce Motivation Lots of Semi structured data Horizontal
Service Integration course. Cassandra
Budapest University of Technology and Economics Department of Measurement and Information Systems Fault Tolerant Systems Research Group Service Integration course Cassandra Oszkár Semeráth Gábor Szárnyas
Comparing SQL and NOSQL databases
COSC 6397 Big Data Analytics Data Formats (II) HBase Edgar Gabriel Spring 2015 Comparing SQL and NOSQL databases Types Development History Data Storage Model SQL One type (SQL database) with minor variations
The Hadoop Eco System Shanghai Data Science Meetup
The Hadoop Eco System Shanghai Data Science Meetup Karthik Rajasethupathy, Christian Kuka 03.11.2015 @Agora Space Overview What is this talk about? Giving an overview of the Hadoop Ecosystem and related
HareDB HBase Client Web Version USER MANUAL HAREDB TEAM
2013 HareDB HBase Client Web Version USER MANUAL HAREDB TEAM Connect to HBase... 2 Connection... 3 Connection Manager... 3 Add a new Connection... 4 Alter Connection... 6 Delete Connection... 6 Clone Connection...
Getting started Cassandra Access control list
Getting started Cassandra Access control list Introduction: This document aims to provide a few easy to follow steps for the first-time user. We will cover the following subjects regarding our access control
Media Upload and Sharing Website using HBASE
A-PDF Merger DEMO : Purchase from www.a-pdf.com to remove the watermark Media Upload and Sharing Website using HBASE Tushar Mahajan Santosh Mukherjee Shubham Mathur Agenda Motivation for the project Introduction
Scaling Up 2 CSE 6242 / CX 4242. Duen Horng (Polo) Chau Georgia Tech. HBase, Hive
CSE 6242 / CX 4242 Scaling Up 2 HBase, Hive Duen Horng (Polo) Chau Georgia Tech Some lectures are partly based on materials by Professors Guy Lebanon, Jeffrey Heer, John Stasko, Christos Faloutsos, Le
Advanced Java Client API
2012 coreservlets.com and Dima May Advanced Java Client API Advanced Topics Originals of slides and source code for examples: http://www.coreservlets.com/hadoop-tutorial/ Also see the customized Hadoop
HBase Java Administrative API
2012 coreservlets.com and Dima May HBase Java Administrative API Originals of slides and source code for examples: http://www.coreservlets.com/hadoop-tutorial/ Also see the customized Hadoop training courses
Overview of Web Services API
1 CHAPTER The Cisco IP Interoperability and Collaboration System (IPICS) 4.5(x) application programming interface (API) provides a web services-based API that enables the management and control of various
Introduction to Hbase Gkavresis Giorgos 1470
Introduction to Hbase Gkavresis Giorgos 1470 Agenda What is Hbase Installation About RDBMS Overview of Hbase Why Hbase instead of RDBMS Architecture of Hbase Hbase interface Summarise What is Hbase Hbase
MariaDB Cassandra interoperability
MariaDB Cassandra interoperability Cassandra Storage Engine in MariaDB Sergei Petrunia Colin Charles Who are we Sergei Petrunia Principal developer of CassandraSE, optimizer developer, formerly from MySQL
Programming Hadoop 5-day, instructor-led BD-106. MapReduce Overview. Hadoop Overview
Programming Hadoop 5-day, instructor-led BD-106 MapReduce Overview The Client Server Processing Pattern Distributed Computing Challenges MapReduce Defined Google's MapReduce The Map Phase of MapReduce
Qsoft Inc www.qsoft-inc.com
Big Data & Hadoop Qsoft Inc www.qsoft-inc.com Course Topics 1 2 3 4 5 6 Week 1: Introduction to Big Data, Hadoop Architecture and HDFS Week 2: Setting up Hadoop Cluster Week 3: MapReduce Part 1 Week 4:
USING HDFS ON DISCOVERY CLUSTER TWO EXAMPLES - test1 and test2
USING HDFS ON DISCOVERY CLUSTER TWO EXAMPLES - test1 and test2 (Using HDFS on Discovery Cluster for Discovery Cluster Users email [email protected] if you have questions or need more clarifications. Nilay
Cloud Scale Distributed Data Storage. Jürmo Mehine
Cloud Scale Distributed Data Storage Jürmo Mehine 2014 Outline Background Relational model Database scaling Keys, values and aggregates The NoSQL landscape Non-relational data models Key-value Document-oriented
Introduction to Big data. Why Big data? Case Studies. Introduction to Hadoop. Understanding Features of Hadoop. Hadoop Architecture.
Big Data Hadoop Administration and Developer Course This course is designed to understand and implement the concepts of Big data and Hadoop. This will cover right from setting up Hadoop environment in
CASSANDRA. Arash Akhlaghi, Badrinath Jayakumar, Wa el Belkasim. Instructor: Dr. Rajshekhar Sunderraman. CSC 8711 Project Report
CASSANDRA Arash Akhlaghi, Badrinath Jayakumar, Wa el Belkasim Instructor: Dr. Rajshekhar Sunderraman CSC 8711 Project Report 1 Introduction The relational model was brought by E.F. Codd s 1970 paper which
Configuration Manual Yahoo Cloud System Benchmark (YCSB) 24-Mar-14 SEECS-NUST Faria Mehak
Configuration Manual Yahoo Cloud System Benchmark (YCSB) 24-Mar-14 SEECS-NUST Faria Mehak Table of Contents 1 Introduction... 3 1.1 Purpose... 3 1.2 Product Information... 3 2 Installation Manual... 3
Integrating Big Data into the Computing Curricula
Integrating Big Data into the Computing Curricula Yasin Silva, Suzanne Dietrich, Jason Reed, Lisa Tsosie Arizona State University http://www.public.asu.edu/~ynsilva/ibigdata/ 1 Overview Motivation Big
Apache HBase. Crazy dances on the elephant back
Apache HBase Crazy dances on the elephant back Roman Nikitchenko, 16.10.2014 YARN 2 FIRST EVER DATA OS 10.000 nodes computer Recent technology changes are focused on higher scale. Better resource usage
Amazon Glacier. Developer Guide API Version 2012-06-01
Amazon Glacier Developer Guide Amazon Glacier: Developer Guide Copyright 2016 Amazon Web Services, Inc. and/or its affiliates. All rights reserved. Amazon's trademarks and trade dress may not be used in
Data storing and data access
Data storing and data access Plan Basic Java API for HBase demo Bulk data loading Hands-on Distributed storage for user files SQL on nosql Summary Basic Java API for HBase import org.apache.hadoop.hbase.*
Big Data and Scripting Systems build on top of Hadoop
Big Data and Scripting Systems build on top of Hadoop 1, 2, Pig/Latin high-level map reduce programming platform interactive execution of map reduce jobs Pig is the name of the system Pig Latin is the
HBase Key Design. 2012 coreservlets.com and Dima May. 2012 coreservlets.com and Dima May
2012 coreservlets.com and Dima May HBase Key Design Originals of slides and source code for examples: http://www.coreservlets.com/hadoop-tutorial/ Also see the customized Hadoop training courses (onsite
Amazon Redshift & Amazon DynamoDB Michael Hanisch, Amazon Web Services Erez Hadas-Sonnenschein, clipkit GmbH Witali Stohler, clipkit GmbH 2014-05-15
Amazon Redshift & Amazon DynamoDB Michael Hanisch, Amazon Web Services Erez Hadas-Sonnenschein, clipkit GmbH Witali Stohler, clipkit GmbH 2014-05-15 2014 Amazon.com, Inc. and its affiliates. All rights
Going Native With Apache Cassandra. QCon London, 2014 www.datastax.com @DataStaxEMEA
Going Native With Apache Cassandra QCon London, 2014 www.datastax.com @DataStaxEMEA About Me Johnny Miller Solutions Architect www.datastax.com @DataStaxEU [email protected] @CyanMiller https://www.linkedin.com/in/johnnymiller
Word count example Abdalrahman Alsaedi
Word count example Abdalrahman Alsaedi To run word count in AWS you have two different ways; either use the already exist WordCount program, or to write your own file. First: Using AWS word count program
Getting Started with SandStorm NoSQL Benchmark
Getting Started with SandStorm NoSQL Benchmark SandStorm is an enterprise performance testing tool for web, mobile, cloud and big data applications. It provides a framework for benchmarking NoSQL, Hadoop,
Easy-Cassandra User Guide
Easy-Cassandra User Guide Document version: 005 1 Summary About Easy-Cassandra...5 Features...5 Java Objects Supported...5 About Versions...6 Version: 1.1.0...6 Version: 1.0.9...6 Version: 1.0.8...6 Version:
Introduction to Big Data Training
Introduction to Big Data Training The quickest way to be introduce with NOSQL/BIG DATA offerings Learn and experience Big Data Solutions including Hadoop HDFS, Map Reduce, NoSQL DBs: Document Based DB
How To Scale Out Of A Nosql Database
Firebird meets NoSQL (Apache HBase) Case Study Firebird Conference 2011 Luxembourg 25.11.2011 26.11.2011 Thomas Steinmaurer DI +43 7236 3343 896 [email protected] www.scch.at Michael Zwick DI
Report Vertiefung, Spring 2013 Constant Interval Extraction using Hadoop
Report Vertiefung, Spring 2013 Constant Interval Extraction using Hadoop Thomas Brenner, 08-928-434 1 Introduction+and+Task+ Temporal databases are databases expanded with a time dimension in order to
t] open source Hadoop Beginner's Guide ij$ data avalanche Garry Turkington Learn how to crunch big data to extract meaning from
Hadoop Beginner's Guide Learn how to crunch big data to extract meaning from data avalanche Garry Turkington [ PUBLISHING t] open source I I community experience distilled ftu\ ij$ BIRMINGHAMMUMBAI ')
NOSQL DATABASES AND CASSANDRA
NOSQL DATABASES AND CASSANDRA Semester Project: Advanced Databases DECEMBER 14, 2015 WANG CAN, EVABRIGHT BERTHA Université Libre de Bruxelles 0 Preface The goal of this report is to introduce the new evolving
Copy the.jar file into the plugins/ subfolder of your Eclipse installation. (e.g., C:\Program Files\Eclipse\plugins)
Beijing Codelab 1 Introduction to the Hadoop Environment Spinnaker Labs, Inc. Contains materials Copyright 2007 University of Washington, licensed under the Creative Commons Attribution 3.0 License --
Elastic Map Reduce. Shadi Khalifa Database Systems Laboratory (DSL) [email protected]
Elastic Map Reduce Shadi Khalifa Database Systems Laboratory (DSL) [email protected] The Amazon Web Services Universe Cross Service Features Management Interface Platform Services Infrastructure Services
Big Data Development CASSANDRA NoSQL Training - Workshop. March 13 to 17-2016 9 am to 5 pm HOTEL DUBAI GRAND DUBAI
Big Data Development CASSANDRA NoSQL Training - Workshop March 13 to 17-2016 9 am to 5 pm HOTEL DUBAI GRAND DUBAI ISIDUS TECH TEAM FZE PO Box 121109 Dubai UAE, email training-coordinator@isidusnet M: +97150
Вовченко Алексей, к.т.н., с.н.с. ВМК МГУ ИПИ РАН
Вовченко Алексей, к.т.н., с.н.с. ВМК МГУ ИПИ РАН Zettabytes Petabytes ABC Sharding A B C Id Fn Ln Addr 1 Fred Jones Liberty, NY 2 John Smith?????? 122+ NoSQL Database
Processing of massive data: MapReduce. 2. Hadoop. New Trends In Distributed Systems MSc Software and Systems
Processing of massive data: MapReduce 2. Hadoop 1 MapReduce Implementations Google were the first that applied MapReduce for big data analysis Their idea was introduced in their seminal paper MapReduce:
Apache Flume and Apache Sqoop Data Ingestion to Apache Hadoop Clusters on VMware vsphere SOLUTION GUIDE
Apache Flume and Apache Sqoop Data Ingestion to Apache Hadoop Clusters on VMware vsphere SOLUTION GUIDE Table of Contents Apache Hadoop Deployment Using VMware vsphere Big Data Extensions.... 3 Big Data
Unleash the Power of HBase Shell
Unleash the Power of HBase Shell Big Data Everywhere Chicago 2014 Jayesh Thakrar [email protected] HBase = Truly Big Data Store..(well, one of many) Proven Scalable Resilient and highly available
MarkLogic Server. MarkLogic Connector for Hadoop Developer s Guide. MarkLogic 8 February, 2015
MarkLogic Connector for Hadoop Developer s Guide 1 MarkLogic 8 February, 2015 Last Revised: 8.0-3, June, 2015 Copyright 2015 MarkLogic Corporation. All rights reserved. Table of Contents Table of Contents
Apache HBase: the Hadoop Database
Apache HBase: the Hadoop Database Yuanru Qian, Andrew Sharp, Jiuling Wang Today we will discuss Apache HBase, the Hadoop Database. HBase is designed specifically for use by Hadoop, and we will define Hadoop
Scaling Up HBase, Hive, Pegasus
CSE 6242 A / CS 4803 DVA Mar 7, 2013 Scaling Up HBase, Hive, Pegasus Duen Horng (Polo) Chau Georgia Tech Some lectures are partly based on materials by Professors Guy Lebanon, Jeffrey Heer, John Stasko,
Hadoop Integration Guide
HP Vertica Analytic Database Software Version: 7.0.x Document Release Date: 2/20/2015 Legal Notices Warranty The only warranties for HP products and services are set forth in the express warranty statements
HBase A Comprehensive Introduction. James Chin, Zikai Wang Monday, March 14, 2011 CS 227 (Topics in Database Management) CIT 367
HBase A Comprehensive Introduction James Chin, Zikai Wang Monday, March 14, 2011 CS 227 (Topics in Database Management) CIT 367 Overview Overview: History Began as project by Powerset to process massive
BERLIN. 2015, Amazon Web Services, Inc. or its affiliates. All rights reserved
BERLIN 2015, Amazon Web Services, Inc. or its affiliates. All rights reserved Build Your Mobile App Faster with AWS Mobile Services Jan Metzner AWS Solutions Architect @janmetzner Danilo Poccia AWS Technical
NoSQL Data Base Basics
NoSQL Data Base Basics Course Notes in Transparency Format Cloud Computing MIRI (CLC-MIRI) UPC Master in Innovation & Research in Informatics Spring- 2013 Jordi Torres, UPC - BSC www.jorditorres.eu HDFS
Chancery SMS 7.5.0 Database Split
TECHNICAL BULLETIN Microsoft SQL Server replication... 1 Transactional replication... 2 Preparing to set up replication... 3 Setting up replication... 4 Quick Reference...11, 2009 Pearson Education, Inc.
Evaluation of NoSQL databases for large-scale decentralized microblogging
Evaluation of NoSQL databases for large-scale decentralized microblogging Cassandra & Couchbase Alexandre Fonseca, Anh Thu Vu, Peter Grman Decentralized Systems - 2nd semester 2012/2013 Universitat Politècnica
OTN Developer Day: Oracle Big Data. Hands On Lab Manual. Introduction to Oracle NoSQL Database
OTN Developer Day: Oracle Big Data Hands On Lab Manual Introduction to ORACLE NOSQL DATABASE HANDS-ON WORKSHOP ii Hands on Workshop Lab Exercise 1 Start and run the Movieplex application. In this lab,
Practical Cassandra. Vitalii Tymchyshyn [email protected] @tivv00
Practical Cassandra NoSQL key-value vs RDBMS why and when Cassandra architecture Cassandra data model Life without joins or HDD space is cheap today Hardware requirements & deployment hints Vitalii Tymchyshyn
Lecture 10: HBase! Claudia Hauff (Web Information Systems)! [email protected]
Big Data Processing, 2014/15 Lecture 10: HBase!! Claudia Hauff (Web Information Systems)! [email protected] 1 Course content Introduction Data streams 1 & 2 The MapReduce paradigm Looking behind the
Hypertable Architecture Overview
WHITE PAPER - MARCH 2012 Hypertable Architecture Overview Hypertable is an open source, scalable NoSQL database modeled after Bigtable, Google s proprietary scalable database. It is written in C++ for
CQL for Cassandra 2.2 & later
CQL for Cassandra 2.2 & later Documentation January 21, 2016 Apache, Apache Cassandra, Apache Hadoop, Hadoop and the eye logo are trademarks of the Apache Software Foundation 2016 DataStax, Inc. All rights
Ankush Cluster Manager - Hadoop2 Technology User Guide
Ankush Cluster Manager - Hadoop2 Technology User Guide Ankush User Manual 1.5 Ankush User s Guide for Hadoop2, Version 1.5 This manual, and the accompanying software and other documentation, is protected
IRF2000 IWL3000 SRC1000 Application Note - Apps with OSGi - Condition Monitoring with WWH push
Version 2.0 Original-Application Note ads-tec GmbH IRF2000 IWL3000 SRC1000 Application Note - Apps with OSGi - Condition Monitoring with WWH push Stand: 28.10.2014 ads-tec GmbH 2014 IRF2000 IWL3000 SRC1000
A programming model in Cloud: MapReduce
A programming model in Cloud: MapReduce Programming model and implementation developed by Google for processing large data sets Users specify a map function to generate a set of intermediate key/value
NoSQL Databases. Institute of Computer Science Databases and Information Systems (DBIS) DB 2, WS 2014/2015
NoSQL Databases Institute of Computer Science Databases and Information Systems (DBIS) DB 2, WS 2014/2015 Database Landscape Source: H. Lim, Y. Han, and S. Babu, How to Fit when No One Size Fits., in CIDR,
Simba Apache Cassandra ODBC Driver
Simba Apache Cassandra ODBC Driver with SQL Connector 2.2.0 Released 2015-11-13 These release notes provide details of enhancements, features, and known issues in Simba Apache Cassandra ODBC Driver with
Managing User Accounts
Managing User Accounts This chapter includes the following sections: Configuring Local Users, page 1 Active Directory, page 2 Viewing User Sessions, page 6 Configuring Local Users Before You Begin You
Big Data and Scripting Systems build on top of Hadoop
Big Data and Scripting Systems build on top of Hadoop 1, 2, Pig/Latin high-level map reduce programming platform Pig is the name of the system Pig Latin is the provided programming language Pig Latin is
HDB++: HIGH AVAILABILITY WITH. l TANGO Meeting l 20 May 2015 l Reynald Bourtembourg
HDB++: HIGH AVAILABILITY WITH Page 1 OVERVIEW What is Cassandra (C*)? Who is using C*? CQL C* architecture Request Coordination Consistency Monitoring tool HDB++ Page 2 OVERVIEW What is Cassandra (C*)?
Project 5 Twitter Analyzer Due: Fri. 2015-12-11 11:59:59 pm
Project 5 Twitter Analyzer Due: Fri. 2015-12-11 11:59:59 pm Goal. In this project you will use Hadoop to build a tool for processing sets of Twitter posts (i.e. tweets) and determining which people, tweets,
Hadoop Project for IDEAL in CS5604
Hadoop Project for IDEAL in CS5604 by Jose Cadena Mengsu Chen Chengyuan Wen {jcadena,mschen,[email protected] Completed as part of the course CS5604: Information storage and retrieval offered by Dr. Edward
Managing User Accounts
Managing User Accounts This chapter includes the following sections: Active Directory, page 1 Configuring Local Users, page 3 Viewing User Sessions, page 5 Active Directory Active Directory is a technology
Cassandra. Jonathan Ellis
Cassandra Jonathan Ellis Motivation Scaling reads to a relational database is hard Scaling writes to a relational database is virtually impossible and when you do, it usually isn't relational anymore The
ITG Software Engineering
Introduction to Apache Hadoop Course ID: Page 1 Last Updated 12/15/2014 Introduction to Apache Hadoop Course Overview: This 5 day course introduces the student to the Hadoop architecture, file system,
Apache HBase NoSQL on Hadoop
Apache HBase NoSQL on Hadoop Project Report Advanced Databases Muhammed Mohiyudheen Ziyad (000423255) Sharma Akshi (000423959) IT4BI Masters - 2015 31/12/2015 Apache HBase NoSQL on Hadoop 1 Table of Contents
Amazon Fulfillment Web Service. Getting Started Guide Version 1.1
Amazon Fulfillment Web Service Getting Started Guide Amazon Fulfillment Web Service: Getting Started Guide Copyright 2010 Amazon Web Services LLC or its affiliates. All rights reserved. Table of Contents
Introduction to Hadoop on the cloud using BigInsights on BlueMix dev@pulse, Feb. 24-25, 2014
Hands on Lab Introduction to Hadoop on the cloud using BigInsights on BlueMix dev@pulse, Feb. 24-25, 2014 Cindy Saracco, Senior Solutions Architect, [email protected], @IBMbigdata Nicolas Morales, Solutions
The Pentaho Big Data Guide
The Pentaho Big Data Guide This document supports Pentaho Business Analytics Suite 4.8 GA and Pentaho Data Integration 4.4 GA, documentation revision October 31, 2012. This document is copyright 2012 Pentaho
HDFS Cluster Installation Automation for TupleWare
HDFS Cluster Installation Automation for TupleWare Xinyi Lu Department of Computer Science Brown University Providence, RI 02912 [email protected] March 26, 2014 Abstract TupleWare[1] is a C++ Framework
Installation Guide. Copyright (c) 2015 The OpenNMS Group, Inc. OpenNMS 17.0.0-SNAPSHOT Last updated 2015-09-22 05:19:20 EDT
Installation Guide Copyright (c) 2015 The OpenNMS Group, Inc. OpenNMS 17.0.0-SNAPSHOT Last updated 2015-09-22 05:19:20 EDT Table of Contents 1. Basic Installation of OpenNMS... 1 1.1. Repositories for
Hadoop Basics with InfoSphere BigInsights
An IBM Proof of Technology Hadoop Basics with InfoSphere BigInsights Unit 2: Using MapReduce An IBM Proof of Technology Catalog Number Copyright IBM Corporation, 2013 US Government Users Restricted Rights
The SkySQL Administration Console
www.skysql.com The SkySQL Administration Console Version 1.1 Draft Documentation Overview The SkySQL Administration Console is a web based application for the administration and monitoring of MySQL 1 databases.
Hadoop Integration Guide
HP Vertica Analytic Database Software Version: 7.1.x Document Release Date: 12/9/2015 Legal Notices Warranty The only warranties for HP products and services are set forth in the express warranty statements
Pro Apache Hadoop. Second Edition. Sameer Wadkar. Madhu Siddalingaiah
Pro Apache Hadoop Second Edition Sameer Wadkar Madhu Siddalingaiah Contents J About the Authors About the Technical Reviewer Acknowledgments Introduction xix xxi xxiii xxv Chapter 1: Motivation for Big
Hadoop Data Warehouse Manual
Ruben Vervaeke & Jonas Lesy 1 Hadoop Data Warehouse Manual To start off, we d like to advise you to read the thesis written about this project before applying any changes to the setup! The thesis can be
Hands-on Exercises with Big Data
Hands-on Exercises with Big Data Lab Sheet 1: Getting Started with MapReduce and Hadoop The aim of this exercise is to learn how to begin creating MapReduce programs using the Hadoop Java framework. In
Benchmarking Top NoSQL Databases Apache Cassandra, Couchbase, HBase, and MongoDB Originally Published: April 13, 2015 Revised: May 27, 2015
Benchmarking Top NoSQL Databases Apache Cassandra, Couchbase, HBase, and MongoDB Originally Published: April 13, 2015 Revised: May 27, 2015 http://www.endpoint.com/ Table of Contents Executive Summary...
Infrastructures for big data
Infrastructures for big data Rasmus Pagh 1 Today s lecture Three technologies for handling big data: MapReduce (Hadoop) BigTable (and descendants) Data stream algorithms Alternatives to (some uses of)
Ankush Cluster Manager - Cassandra Technology User Guide
Ankush Cluster Manager - Cassandra Technology User Guide Ankush User s Guide for Cassandra, Version 1.5 This manual, and the accompanying software and other documentation, is protected by U.S. and international
Enabling SOX Compliance on DataStax Enterprise
Enabling SOX Compliance on DataStax Enterprise Table of Contents Table of Contents... 2 Introduction... 3 SOX Compliance and Requirements... 3 Who Must Comply with SOX?... 3 SOX Goals and Objectives...
Integrating VoltDB with Hadoop
The NewSQL database you ll never outgrow Integrating with Hadoop Hadoop is an open source framework for managing and manipulating massive volumes of data. is an database for handling high velocity data.
HBase. Lecture BigData Analytics. Julian M. Kunkel. [email protected]. University of Hamburg / German Climate Computing Center (DKRZ)
HBase Lecture BigData Analytics Julian M. Kunkel [email protected] University of Hamburg / German Climate Computing Center (DKRZ) 11-12-2015 Outline 1 Introduction 2 Excursion: ZooKeeper 3 Architecture
Map Reduce & Hadoop Recommended Text:
Big Data Map Reduce & Hadoop Recommended Text:! Large datasets are becoming more common The New York Stock Exchange generates about one terabyte of new trade data per day. Facebook hosts approximately
MADOCA II Data Logging System Using NoSQL Database for SPring-8
MADOCA II Data Logging System Using NoSQL Database for SPring-8 A.Yamashita and M.Kago SPring-8/JASRI, Japan NoSQL WED3O03 OR: How I Learned to Stop Worrying and Love Cassandra Outline SPring-8 logging
Hands on exercise for
Hands on exercise for João Miguel Pereira 2011 0 Prerequisites, assumptions and notes Have Maven 2 installed in your computer Have Eclipse installed in your computer (Recommended: Indigo Version) I m assuming
WebSphere Business Monitor V7.0: Clustering Single cluster deployment environment pattern
Copyright IBM Corporation 2010 All rights reserved WebSphere Business Monitor V7.0: Clustering Single cluster deployment environment pattern What this exercise is about... 2 Exercise requirements... 2
