IPRO ecapture Performance Report using BlueArc Titan Network Storage System



Similar documents
A High-Performance Storage and Ultra-High-Speed File Transfer Solution

Building a Successful Strategy To Manage Data Growth

ediscovery Journal Report: Digital Reef & BlueArc ediscovery Software Performance Benchmark Test

Application Brief: Using Titan for MS SQL

Best Practices for Deploying SSDs in a Microsoft SQL Server 2008 OLTP Environment with Dell EqualLogic PS-Series Arrays

MaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products

Using VMware VMotion with Oracle Database and EMC CLARiiON Storage Systems

The Evolution of Microsoft SQL Server: The right time for Violin flash Memory Arrays

8Gb Fibre Channel Adapter of Choice in Microsoft Hyper-V Environments

Integrated Grid Solutions. and Greenplum

Milestone Solution Partner IT Infrastructure MTP Certification Report Scality RING Software-Defined Storage

EMC Unified Storage for Microsoft SQL Server 2008

Oracle Database Scalability in VMware ESX VMware ESX 3.5

White Paper. Recording Server Virtualization

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

How To Test For Performance And Scalability On A Server With A Multi-Core Computer (For A Large Server)

Accelerating Server Storage Performance on Lenovo ThinkServer

WHITE PAPER The Storage Holy Grail: Decoupling Performance from Capacity

Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study

Accelerating Microsoft Exchange Servers with I/O Caching

Big data management with IBM General Parallel File System

Business white paper. HP Process Automation. Version 7.0. Server performance

WHITE PAPER Optimizing Virtual Platform Disk Performance

BlueArc unified network storage systems 7th TF-Storage Meeting. Scale Bigger, Store Smarter, Accelerate Everything

Dell Compellent Storage Center SAN & VMware View 1,000 Desktop Reference Architecture. Dell Compellent Product Specialist Team

Solving I/O Bottlenecks to Enable Superior Cloud Efficiency

Using Data Domain Storage with Symantec Enterprise Vault 8. White Paper. Michael McLaughlin Data Domain Technical Marketing

Comparison of Hybrid Flash Storage System Performance

Virtualizing SQL Server 2008 Using EMC VNX Series and Microsoft Windows Server 2008 R2 Hyper-V. Reference Architecture

EMC XTREMIO EXECUTIVE OVERVIEW

DELL TM PowerEdge TM T Mailbox Resiliency Exchange 2010 Storage Solution

Dell Virtualization Solution for Microsoft SQL Server 2012 using PowerEdge R820

Performance characterization report for Microsoft Hyper-V R2 on HP StorageWorks P4500 SAN storage

Pivot3 Reference Architecture for VMware View Version 1.03

Performance Characteristics of VMFS and RDM VMware ESX Server 3.0.1

Scala Storage Scale-Out Clustered Storage White Paper

EMC Backup and Recovery for Microsoft Exchange 2007 SP2

EMC Business Continuity for Microsoft SQL Server 2008

Dragon Medical Enterprise Network Edition Technical Note: Requirements for DMENE Networks with virtual servers

Clustering Windows File Servers for Enterprise Scale and High Availability

Cisco UCS and Fusion- io take Big Data workloads to extreme performance in a small footprint: A case study with Oracle NoSQL database

Microsoft SharePoint Server 2010

LSI MegaRAID CacheCade Performance Evaluation in a Web Server Environment

Navisphere Quality of Service Manager (NQM) Applied Technology

HP reference configuration for entry-level SAS Grid Manager solutions

Introduction. Scalable File-Serving Using External Storage

Toolbox 4.3. System Requirements

Benchmarking Microsoft SQL Server Using VMware ESX Server 3.5

EMC SOLUTION FOR SPLUNK

SMB Direct for SQL Server and Private Cloud

Reference Architecture for Dell VIS Self-Service Creator and VMware vsphere 4

Performance and scalability of a large OLTP workload

NEWSTAR ENTERPRISE and NEWSTAR Sales System Requirements

WHITE PAPER Guide to 50% Faster VMs No Hardware Required

Amadeus SAS Specialists Prove Fusion iomemory a Superior Analysis Accelerator

Analysis of VDI Storage Performance During Bootstorm

W H I T E P A P E R. Best Practices for High Performance NFS Storage with VMware

Estimate Performance and Capacity Requirements for Workflow in SharePoint Server 2010

Evaluation Report: Accelerating SQL Server Database Performance with the Lenovo Storage S3200 SAN Array

Cisco, Citrix, Microsoft, and NetApp Deliver Simplified High-Performance Infrastructure for Virtual Desktops

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

Increasing Storage Performance, Reducing Cost and Simplifying Management for VDI Deployments

EMC Virtual Infrastructure for Microsoft Applications Data Center Solution

Technical White Paper. Symantec Backup Exec 10d System Sizing. Best Practices For Optimizing Performance of the Continuous Protection Server

Hardware Configuration Guide

EMC Virtual Infrastructure for SAP Enabled by EMC Symmetrix with Auto-provisioning Groups, Symmetrix Management Console, and VMware vcenter Converter

Scalability. Microsoft Dynamics GP Benchmark Performance: Advantages of Microsoft SQL Server 2008 with Compression.

Windows 8 SMB 2.2 File Sharing Performance

Cisco Unified Computing System and EMC VNXe3300 Unified Storage System

Evaluation of Dell PowerEdge VRTX Shared PERC8 in Failover Scenario

Netwrix Auditor for SQL Server

Improve application performance and scalability with Adobe ColdFusion 9

Using Synology SSD Technology to Enhance System Performance Synology Inc.

BEST PRACTICES FOR INTEGRATING TELESTREAM VANTAGE WITH EMC ISILON ONEFS

Leveraging EMC Fully Automated Storage Tiering (FAST) and FAST Cache for SQL Server Enterprise Deployments

Virtual Desktop Infrastructure (VDI) made Easy

Avid ISIS

RFP-MM Enterprise Storage Addendum 1

Boost Database Performance with the Cisco UCS Storage Accelerator

Removing Performance Bottlenecks in Databases with Red Hat Enterprise Linux and Violin Memory Flash Storage Arrays. Red Hat Performance Engineering

Microsoft Exchange Server 2003 Deployment Considerations

Doubling the I/O Performance of VMware vsphere 4.1

Installation Guide July 2009

HP SN1000E 16 Gb Fibre Channel HBA Evaluation

Gridstore Proves Storage Resiliency

WHITE PAPER Guide to 50% Faster VMs No Hardware Required

WHITEPAPER: Understanding Pillar Axiom Data Protection Options

Performance Test Report: Novell iprint Appliance 1.1

Virtualizing SQL Server 2008 Using EMC VNX Series and Microsoft Windows Server 2008 R2 Hyper-V. Proven Solution Guide

High Performance Server SAN using Micron M500DC SSDs and Sanbolic Software

Data Sheet: Archiving Symantec Enterprise Vault Discovery Accelerator Accelerate e-discovery and simplify review

Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage

Litigation Solutions. insightful interactive culling. distributed ediscovery processing. powering digital review

Netwrix Auditor for Active Directory

Enterprise Storage Solution for Hyper-V Private Cloud and VDI Deployments using Sanbolic s Melio Cloud Software Suite April 2011

Zeus Traffic Manager VA Performance on vsphere 4

Best Practices for Optimizing SQL Server Database Performance with the LSI WarpDrive Acceleration Card

EMC Virtual Infrastructure for Microsoft Applications Data Center Solution

ADVANCED NETWORK CONFIGURATION GUIDE

Virtualized Exchange 2007 Archiving with EMC Xtender/DiskXtender to EMC Centera

Transcription:

IPRO ecapture Performance Report using BlueArc Titan Network Storage System Allen Yen, BlueArc Corp Jesse Abrahams, IPRO Tech, Inc

Introduction IPRO ecapture is an e-discovery application designed to handle large projects with speed and scalability. It is a SQL based, client-server application with multi-threading capability that can be distributed across any number of workstations. IPRO refers to these distributed workstations as Workers. All processing is managed from a central console and all processing operations are performed on workers without operator intervention. BlueArc s Titan Network Storage System represents an evolutionary step in file servers by creating a hardware-based file system that can scale bandwidth, IOPS, and overall data capacity well beyond conventional software-based devices. With its ability to virtualize a massive storage pool of up to four usable petabytes of tiered storage, Titan can scale with growing data requirements, offering a competitive advantage for businesses, researchers, or other enterprises seeking to better manage data growth while still ensuring optimal performance. The goals of this benchmark are to measure the workload characteristics and performance benchmarks of the Titan storage system in a processing architecture including IPRO ecapture running on a pool of Workers, and to relate Titan performance results with the workflow of IPRO ecapture. This benchmark will also measure how IPRO ecapture and Titan respond in a variety of scalability scenarios. 1. Data Set & Test Equipment 1.1 Data Set No standard data set for testing of electronic discovery processing solutions currently exists. There are industry heuristics that indicate what types of files are used in typical forensic data collections, typical data sets are based on how many pages per document per document type are typical, how many documents per gigabyte, how many gigabytes can be processed per day, etc. None of this is well defined. The lack of standard data sets that can be used to evaluate various systems and software architectures was a principal motivator for the development of this benchmark. The distribution of file types in the dataset was jointly defined by Dutton, LLC and IPRO Tech and was then generated by IPRO. This data set is designed to be representative of collected ESI in a moderate-size electronic discovery project (45 Gigabytes) and was designed to yield approximately 1 million items and over 3 million pages. The data set was organized into 9 custodian directories of approximately 4-6 GB each. This benchmark uses this data set exclusively throughout. 1.2 Test Equipment A summary of hardware and software used for this benchmarking follows: PAGE 1

Hardware IPRO ecapture Worker One Dell R9 Server One Dell R9 Server Two Dell 295 One Supermicro 2U Server SQL Server One Dell 29 One Dell 295 Storage Titan 321 Storage System Software Four Quad-core Intel CPU at 2.33 GHz 64 GB RAM Four Quad-core Intel CPU at 2.33 GHz 24 GB RAM Two Dual-Core Intel CPU at 2. GHz 8 GB RAM Two Dual-Core Intel CPU at 2.5 GHz 16 GB RAM Dual Quad-Core Intel CPU at 3.16 GHz 32 GB RAM Dual Core Intel CPU at 2. GHz 24 GB RAM 2ea RC16 arrays, each with 64 3GB 15K disks ESX Server 3.5 Application Worker Window 23 Server SP1 32 bit 2 VCPU 2 GB RAM SQL Server SQL Server 23 x64 SP2 SQL RAM 28GB 2 Benchmark Methodology The application used for this benchmark was the e-discovery software IPRO ecapture, version 3.3.. The primary features of the software are divided into four job types: Discovery, Data Extraction, Processing and Exporting. To understand workload characteristics for each job, all four jobs were tested and performance numbers collected. Based on these tests, Discovery jobs were chosen as the most I/O-intensive jobs that would put the most stress on the storage environment, and were therefore used for the entire benchmark test. Two approaches were used to benchmark the performance of IPRO ecapture and the Titan storage system: scale up and scale out. The scale-up approach utilized additional IPRO Workers to increase the workload and performance on a single SQL database. The scale-out approach utilized additional SQL servers and databases to distribute more data and tasks across all IPRO Workers while at the same time increasing the workload on the Titan. 3.1 Software configuration IPRO ecapture is an e-discovery platform that uses a client-server model comprised of the following modules: Controller Job creation, preparation and monitoring Queue Manager Task distribution and coordination amongst all connected Workers Worker Performs all required work as determined by task definitions. Examples of such work include: extracting metadata and text, converting electronic files to TIFF, JPG, or PDF, and creating word index files used for keyword searching. SQL Server Primary communication method between all IPRO ecapture components, and manages SQL databases that store application settings, file information and file relationships. Storage Used to access discovery directories and store application output (extracted files, indexes, images and text). The hardware and software outlined above support the modules either directly (the software is installed on the hardware) or indirectly (the software uses the hardware to function normally). PAGE 2

The primary functions being performed during the Discovery portion of IPRO ecapture (a Discovery Job ) are: Discover all files in specified directories, including subdirectories. Extract embedded items (if necessary) from mail stores, email messages, archives, and other compound documents; put the files in a separate location on disk and record their document relations in the database. Index the contents and metadata of each item, for the purposes of filtering on responsive search terms; indexes are stored on disk. The Discovery portion of IPRO ecapture performs several operations designed to facilitate and increase speed in other aspects of the processing workflow. Conversion to tiff and export are enabled for distributed processing due to the groundwork operations performed within the discovery phase. 3.2 Storage configuration The following is the configuration of the Titan storage system: Each RAID group is RAID5, 4 data and one parity drive, with 64K block size. One hot spare per enclosure. Three storage spools were created: - App_pool: 12 RAID groups (Titan system drives). One 9 GB file system was created to serve application data, which was expanded as necessary. - SQLlog: 4 system drives. 2 iscsi LUN were created for database logging. The size of each log was 15G and 2G each. - SQLdata: 8 system drives. 4 iscsi LUN were created for base tables and indexing. The size of each LUN was 4G, 5G, 15G and 5GB each. Data Networking Two aggregation port groups were created. Group One had 3ea Gigabit Ethernet ports configured as a VLAN to serve the application data. Group Two had another VLAN configured with 3ea Gigabit Ethernet ports to serve SQL databases and logs. 3 Benchmark Results For each benchmark test, the Discovery feature of IPRO ecapture was performed on the same data set. To keep accurate comparisons, for every benchmark job fresh databases were created to ensure no pre-existing data in the databases caused additional overhead. 4.1 Scale-Up: Single Database with Variable Number of IPRO Workers Total GB (approx) Worker Count Elapsed Time Before After GByte/HR 1W 6:24:55 45 75 11.5 3W 4:45:36 45 75 15.8 Table 1: Each Worker has four threads running at the same time. Data set size increased from initial 45GB to 75GB at the completion of the Discovery job. PAGE 3

IPRO Discovery Process, CIFS/sec 1 Worker 3 Worker 4 3 CIFS/sec 2 1 1 2 3 4 5 Time ( 7 hours) Figure 1 shows the results of scalability from 1 and 3 Workers. In Figure 1, the data load from the left side of the chart started at approximately 39, OP/sec with 3 Workers. It lasted for 3 minutes (1 Workers took longer to complete) and slowed down to 2, OP/sec in the middle of the job, then down to 1, OP/sec, eventually dropping to zero when the job was complete. The Discovery function of IPRO ecapture is generally more I/O and CPU intensive on storage devices because existing files are continuously being accessed simultaneous to new files and folders being created and accessed. Additionally, after primary discovery functions complete the files are accessed again to create searchable indexes that are also stored on disk (see Section 3.1). The Data Extract, Processing, and Export functions of IPRO ecapture are more CPU intensive on the IPRO Worker because image and text files are created locally and then copied to a network location on the storage device. These functions are not as I/O intensive on the Titan because files are simply copied from the temporary local location to the network location. Based on the results shown in Figure 1, it is clear the Titan is capable of handling more than 1 IPRO Workers for this specific data set, reaching only 27, OP/sec in performance with this number of multi-threaded Workers. For 3 IPRO Workers, the maximum OP/sec Titan achieved was around 39, OP/sec when CPU utilization was up to 1%. Additional IPRO Workers also raised the through-put rate from 11.5 to 15.8 gigabytes per hour, as shown in Table 1. Finally, additional Workers and increased Titan activity led to a faster completion time for the 3-Worker job. PAGE 4

IPRO Discovery Process, SQL Server, iscsi/sec 1 Worker 3 Worker 12 9 iscsi 6 3 1 2 3 4 5 Time ( 7 hours) Figure 2: SQL Server database activity. Figure 2 shows the overall SQL server activity including I/O for the database and log files. An iscsi LUN was created and used by the SQL server to access the databases. It is important to know any potential bottlenecks in the database as that may degrade performance, especially for SQL logging which is a single thread sequential write workload. With 1 IPRO Workers (slowest completion time), the peak I/O activity started late and completed last. With 3 IPRO Workers, iscsi OP/sec peaked at over 1, dramatically decreasing as primary discovery tasks completed; SQL activity remained low for the remainder of the job as indexing tasks completed. 4.2 Scale-Out: Multiple Databases with Variable Number of IPRO Workers Total GB (approx) DB Instance Worker Elapse Time Before After GByte/HR 1 3 4:45:35 45 75 15.8 2 3 6:22:14 9 15 23.6 Table 2: Elapsed time between one and two database instances. Each IPRO Worker has four threads running at the same time. For the two database instance, the same data set was run simultaneously on the two databases (two jobs at 45 GB each) for a total of 9 GB. PAGE 5

One vs. Two DB Instance 45 One DB Two DB 4 35 3 25 OP/sec 2 15 1 5 1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 31 33 35 37 39 41 43 45 47 Time ( 7 hours) Figure 3. 3 Workers for one database, 3 Workers for two databases. 3 Starting Size Final Size GB/hour 25 23.6 2 Gigabytes 15 15.8 15 1 75 5 9 45 One DB Two DB Figure 4. Starting (compressed) and Final (uncompressed) data set sizes for one- and two-database tests, and the calculated GByte/hour from Table 2. Table 2 shows the elapsed time to completion for Discovery Jobs run on two SQL servers (one database on each SQL server). Both jobs were run simultaneously, with the longest-running job finishing in less than 6.5 hours. A total of 9 GB of data, twice the original amount of data, was discovered with the same number of Workers. Figure 4 shows the original size and the final size of the data sets for each job. These results show that in only 5% more time over the single-database job (less than 2 additional hours), twice the amount of data was discovered. A linear increase would have yielded over 8 elapsed hours instead of 6.5 (actual) elapsed hours. The results also show that data was discovered at a higher rate per hour during the two-database job (see Figure 4). In addition, Figure 3 shows that Titan I/O peaked during the initial discovery tasks and remained high 1.5 times longer than the I/O produced on a single database (3 min for the single database, 45 min for two databases). This is due to duplicate data sets and database activity producing twice the workload on the Titan. PAGE 6

1. Multiple IPRO ecapture jobs running simultaneously on one or many SQL servers and/or databases is a common realistic scenario; see Section 6 for details. More processing resources on the Titan became available as the Discovery jobs moved toward completion since OP/sec decreases over the duration of the job. Due to hardware availability constraints additional IPRO Workers and SQL servers could not be added for these benchmark tests. Based on the above results, however, the Titan is capable of handling additional IPRO Workers and additional SQL servers and database instances for long periods of sustained activity without seeing any performance degradation. IPRO ecapture performance is not affected by this increased activity as well; according to the results, a set number of Workers processed more data using two database instances, faster than expected 1. 4.3 Titan Utilization and Disk Read/Write Latency Titan Resources, Disk Read/Write Latency 12 FSB FSA Disk-R Disk-W OP/sec 45 1 4 35 8 3 Latency 6 25 2 OP/sec 4 15 1 2 5 1 4 7 1 13 16 19 22 25 28 31 34 37 4 43 46 49 52 55 58 61 64 67 7 73 76 79 82 85 88 91 Figure 5: Titan OP/sec, CPU utilization and Disk latency performance on one database with 3 IPRO Workers. Figure 5 above consists of snapshots from different stages of the 3 Worker test run on a single SQL server and database. The snapshots were collected at stages of the job when OP/sec was at 39K, 26K and 18K OP/sec; Figure 4 shows the results of each variable and how they relate to each other. When OP/sec was in the 39K range, the Titan FSA and FSB were 99% and 6% utilized. When OP/sec was reduced to 26K, both the FSA and FSB were down to the mid 5% and the mid 4% range. At the end of benchmark run, FSB had higher utilization then FSA because of indexing tasks which require more resources of FSB. In addition to resource utilization, disk I/O latency is also important to measure and probably is the most critical in the entire data path. When a read operation was cache-missed in the Titan or storage array controller, it will be accessed from disk which involves a mechanical operation and will degrade performance. When I/O is write-missed, it has to wait for all the dirty data de-staging to the disk before operation can be performed. When the FSA and FSB resources were up to 1%, it is possible that the resources were busy waiting for I/O to complete. However, when disk latency was in the optimal range: read latency was less than 5 milliseconds and write latency was no more than 25 milliseconds throughout the entire benchmark measurement, regardless of CPU utilization. These numbers indicate that IPRO ecapture was not disk bound, and that the Titan is capable of receiving more I/O by adding additional front end load, such as more IPRO Workers and additional SQL servers. PAGE 7

4.4 Command queuing and queue depth on Titan Command tag queuing is also an important variable in benchmark activity. When commands are queued on the Titan, it has to wait for storage resource availability before it can process the request from the client. In the entire benchmark testing, there was not a high number of commands queuing from the Titan to the storage array. This shows that disk latency is not a bottleneck in this test. 5. Conclusion The results in this benchmark study show that the Titan storage system is capable of handling extended periods of high activity and workload resulting from running the IPRO ecapture application on an array of Workers. The disk read/write latency is not affected when Titan resource utilization reached 1%, and the maximum performance was up to 4 OP/sec for the 3-45 minutes of peak time. Results also show that there would be potential for further expansion by either adding more Workers or adding additional SQL servers and/or database instances, all without any performance degradation on the Titan. 6. Notes In real-world scenarios, multiple Discovery Jobs are run simultaneously on one system, discovering data sets of a size the same or greater than what was discovered for these benchmark tests (45 GB and 9 GB compressed). This, in addition to Process Jobs, Data Extract Jobs and Export Jobs (see Section 3) also running simultaneously, would produce consistent activity and workload on the Titan storage system, as well as increased data through-put per hour through the IPRO ecapture system. Based on the conclusions above, such load would also not negatively affect performance on the Titan. It is also common to have a higher number of IPRO Workers per system, with the largest IPRO ecapture customers implementing 5 to 1 IPRO Workers across a single SQL server or multiple SQL servers. Additional IPRO Workers, as shown in Section 4.1, increases activity on the Titan as well. Improvements to the IPRO ecapture task distribution model, available in the new release of IPRO ecapture (not used for these benchmark tests), will allow the user to make adjustments to task distribution settings to reflect the number of available IPRO Workers connected to the system; this in turn will improve completion times for all job types. 7. Recommendations A good deal of time was spent configuring a bare-metal Worker environment, then a virtualized Worker environment. It is recommended that the customer properly configure the SQL server and storage system to the potential workload that they intend to run. With these important aspects of the environment set, time can be spent configuring IPRO ecapture to run IPRO Workers on bare-metal machines (a traditional configuration) or in a virtualized environment. If virtualization is necessary, it is recommended that separate benchmark tests be performed to determine what impact (if any) a shift from bare-metal to virtualization may have on performance. There are many potential solutions for scaling IPRO ecapture. Using each individual solution effectively can be complicated. Real benefits are gained when different technologies are combined together to produce a high performance solution for the customer. PAGE 8

About BlueArc BlueArc is a leading provider of high performance unified network storage systems to enterprise markets, as well as data intensive markets, such as electronic discovery, entertainment, federal government, higher education, Internet services, oil and gas and life sciences. Our products support both network attached storage, or NAS, and storage area network, or SAN, services on a converged network storage platform. We enable companies to expand the ways they explore, discover, research, create, process and innovate in data-intensive environments. Our products replace complex and performance-limited products with high performance, scalable and easy to use systems capable of handling the most data intensive applications and environments. Further, we believe that our energy efficient design and our products ability to consolidate legacy storage infrastructures, dramatically increases storage utilization rates and reduces our customers total cost of ownership. About IPRO Tech IPRO offers a diverse suite of superior litigation software tools for e-discovery, database management, review and production and a complete line of products for both workgroups and high-volume production environments. IPRO has thousands of systems of various sizes in law firms, corporate and government legal departments, and litigation service organizations. Millions of documents are processed using IPRO software annually. The American Lawyer Magazine recently completed a survey of the top 1 law firms in the United States and found that IPRO software is by far the most widely used litigation document software. To learn more about IPRO software tools, visit www.iprotech.com or call 888-477-6463. BlueArc Corporation Corporate Headquarters 5 Rio Robles Drive San Jose, CA 95134 t 48 576 66 f 48 576 661 www.bluearc.com BlueArc UK Ltd. European Headquarters Queensgate House Cookham Road Bracknell RG12 1RB, United Kingdom t +44 () 1344 48 2 f +44 () 1344 48 22 29 BlueArc Corporation. All rights reserved. The BlueArc logo is a registered trademark of BlueArc Corporation. 1/9 WP-IPRO-1