www.progress.com DEPLOYMENT ARCHITECTURE FOR JAVA ENVIRONMENTS

Similar documents
DEPLOYMENT ARCHITECTURE FOR MICROSOFT.NET ENVIRONMENTS

I D C V E N D O R S P O T L I G H T. C o r t i c o n T e chnologies: A S o l u t i o n t o

IBM WebSphere ILOG Rules for.net

Service Mediation. The Role of an Enterprise Service Bus in an SOA

PROGRESS DATADIRECT QA AND PERFORMANCE TESTING EXTENSIVE TESTING ENSURES DATA CONNECTIVITY THAT WORKS

WEB SERVICES. Revised 9/29/2015

1 What Are Web Services?

A standards-based approach to application integration

CACHÉ: FLEXIBLE, HIGH-PERFORMANCE PERSISTENCE FOR JAVA APPLICATIONS

1 What Are Web Services?

Open source business rules management system

Version Overview. Business value

BUILDING FLEXIBLE ENTERPRISE PROCESSES USING ORACLE BUSINESS RULES AND BPEL PROCESS MANAGER. An Oracle White Paper Jan 2005

CA Workload Automation Agents Operating System, ERP, Database, Application Services and Web Services

Converting Java EE Applications into OSGi Applications

Open Source Business Rules Management System Enables Active Decisions

Reusing Existing * Java EE Applications from Oracle SOA Suite

Motivation Definitions EAI Architectures Elements Integration Technologies. Part I. EAI: Foundations, Concepts, and Architectures

Deploying Rule Applications

applications. JBoss Enterprise Application Platform

Developing Java Web Services

IBM Rational Asset Manager

BEA AquaLogic Integrator Agile integration for the Enterprise Build, Connect, Re-use

C05 Discovery of Enterprise zsystems Assets for API Management

An Oracle White Paper February Oracle Data Integrator 12c Architecture Overview

Increasing IT flexibility with IBM WebSphere ESB software.

JReport Server Deployment Scenarios

Jitterbit Technical Overview : Microsoft Dynamics CRM

Orchestrating Web Services: The Case for a BPEL Server. An Oracle White Paper June 2004

ILOG JRules Performance Analysis and Capacity Planning

An introduction to SOA and the HP NonStop server environment

How To Use X Query For Data Collection

BUSINESS INTELLIGENCE ANALYTICS

IBM InfoSphere MDM Server v9.0. Version: Demo. Page <<1/11>>

How To Create A C++ Web Service

SONIC ESB 7. KEY CAPABILITIES > Connects, mediates and controls. KEY BENEFITS > Creates new processes using

What You Need to Know About Transitioning to SOA

Copyright 2012, Oracle and/or its affiliates. All rights reserved.

Exam Name: IBM InfoSphere MDM Server v9.0

What is the NXTware Evolution Server Peter Marquez, Product Marketing ecube Systems

IBM WebSphere Operational Decision Management Improve business outcomes with real-time, intelligent decision automation

JVA-561. Developing SOAP Web Services in Java

CA IDMS Server r17. Product Overview. Business Value. Delivery Approach

Introduction to WebSphere Process Server and WebSphere Enterprise Service Bus

CERTIFIED MULESOFT DEVELOPER EXAM. Preparation Guide

SOACertifiedProfessional.Braindumps.S90-03A.v by.JANET.100q. Exam Code: S90-03A. Exam Name: SOA Design & Architecture

SERVICE-ORIENTED MODELING FRAMEWORK (SOMF ) SERVICE-ORIENTED SOFTWARE ARCHITECTURE MODEL LANGUAGE SPECIFICATIONS

CA Workload Automation Agents for Mainframe-Hosted Implementations

CICS Modernization & Integration

a division of Technical Overview Xenos Enterprise Server 2.0

PASS4TEST 専 門 IT 認 証 試 験 問 題 集 提 供 者

SAVVION MANAGEMENT SYSTEM

Service Computing: Basics Monica Scannapieco

SCA-based Enterprise Service Bus WebSphere ESB

The IBM Cognos Platform for Enterprise Business Intelligence

Inside the Digital Commerce Engine. The architecture and deployment of the Elastic Path Digital Commerce Engine

Developing SOA solutions using IBM SOA Foundation

JBoss enterprise soa platform

PROGRESS OPENEDGE PRO2 DATA REPLICATION

ITS. Java WebService. ITS Data-Solutions Pvt Ltd BENEFITS OF ATTENDANCE:

AquaLogic ESB Design and Integration (3 Days)

VALLIAMMAI ENGINEERING COLLEGE SRM NAGAR, KATTANKULATHUR DEPARTMENT OF COMPUTER APPLICATIONS SUBJECT : MC7502 SERVICE ORIENTED ARCHITECTURE

CA Workload Automation

IBM Rational Web Developer for WebSphere Software Version 6.0

White Paper: 1) Architecture Objectives: The primary objective of this architecture is to meet the. 2) Architecture Explanation

JBOSS ENTERPRISE APPLICATION PLATFORM MIGRATION GUIDELINES

MD Link Integration MDI Solutions Limited

LinuxWorld Conference & Expo Server Farms and XML Web Services

Service-Oriented Architecture and Software Engineering

Business Process Management Enabled by SOA

High Availability for Citrix XenApp

Overview: Siebel Enterprise Application Integration. Siebel Innovation Pack 2013 Version 8.1/8.2 September 2013

An Oracle White Paper October Oracle Data Integrator 12c New Features Overview

The IBM Cognos Platform

Answers to Top BRMS Questions

David Pilling Director of Applications and Development

Emerging Technologies Shaping the Future of Data Warehouses & Business Intelligence

Service Oriented Architecture

Service-Oriented Architecture and its Implications for Software Life Cycle Activities

IBM Rational Rapid Developer Components & Web Services

What can DDS do for You? Learn how dynamic publish-subscribe messaging can improve the flexibility and scalability of your applications.

OpenText Output Transformation Server

Copyright. Restricted Rights Legend. Trademarks or Service Marks. Copyright 2003 BEA Systems, Inc. All Rights Reserved.

Oracle Service Bus Examples and Tutorials

Complementing Your Web Services Strategy with Verastream Host Integrator

FUSE-ESB4 An open-source OSGi based platform for EAI and SOA

SOA Fundamentals For Java Developers. Alexander Ulanov, System Architect Odessa, 30 September 2008

What Is the Java TM 2 Platform, Enterprise Edition?

Service Oriented Architecture 1 COMPILED BY BJ

Service-Oriented Integration: Managed File Transfer within an SOA (Service- Oriented Architecture)

What is it? What does it do? Benefits

SOA REFERENCE ARCHITECTURE: WEB TIER

Choose an IBM WebSphere Application Server configuration to suit your business needs

Business Process Management Tampereen Teknillinen Yliopisto

Best Practices for Implementing High Availability for SAS 9.4

> ACCESSING SQL SERVER FROM IBM COGNOS BI SERVER.

Java EE 7: Back-End Server Application Development

Transcription:

DEPLOYMENT ARCHITECTURE FOR JAVA ENVIRONMENTS

TABLE OF CONTENTS Introduction 1 Progress Corticon Product Architecture 1 Deployment Options 2 Invoking Corticon Decision Services 4 Corticon Rule Engine 5 Enterprise Data Connectivity 6 Clustering and Load Balancing 6 Mainframe Interoperability 7 Summary 8

1 INTRODUCTION The Progress Corticon Business Rules Management System ( Corticon BRMS ) fits naturally in today s service-oriented architectures (SOA), deploying as a service, and leverages the enterprise-class performance, scalability, and high-availability features of leading application servers. It has been designed to support standards critical to enterprise customers building and integrating composite applications. The Progress Corticon Server is the patented no-coding rules engine for Corticon BRMS, with performance unmatched in the industry. This document describes Corticon Server s JAVA runtime architecture with particular emphasis on how Corticon Server fits within various enterprise environments. PROGRESS CORTICON PRODUCT ARCHITECTURE The Corticon BRMS consists of a set of rule modeling tools ( Corticon Studio ) and deployment tools ( Corticon Deployment Console ), along with a highly scalable execution environment ( Corticon Server ). The business rules are modeled as rule sets and deployed as Corticon decision services. Client applications call decision services to apply decision-making logic (i.e., business rules) as needed within the application flow. These fully encapsulated decision services can be invoked as web services or in-process, depending on technical requirements. A single instance of Corticon Server can supply decision services to diverse client applications. Figure 1 illustrates the Corticon BRMS architecture for Java environments. PROGRESS CORTICON STUDIO PROGRESS CORTICON DEPLOYMENT CONSOLE JAVA APPLICATION SERVER (JEE) PROGRESS CORTICON SERVER ENTERPRISE DATA CONNECTOR JAVA VIRTUAL MACHINE OPERATING SYSTEM WEB SERVICE Figure 1 Corticon product architecture for Java environments POLICY MAKERS BUSINESS AND IT ENTERPRISE APPLICATIONS ENTERPRISE DATA SOURCES

2 Consider the following: 1. Each decision service encapsulates the logic of a single decision-making activity (e.g., calculate a rate, verify a claim, accept an applicant). Corticon Studio provides tools to analyze and test the logic, ensuring that it is complete and unambiguous; it always provides one, and only one, correct answer. 2. Corticon Server may be deployed in a Java EE web/application server or may be called inprocess from a Java client application. 3. Decision services may be deployed into Corticon Server using a Corticon Deployment Descriptor ( CDD ) or via Corticon Server APIs. All necessary deployment artifacts, including the executable services and WSDLs, can be generated by Corticon tools. 4. Client applications include any business program that uses decision services or business process management (BPM) systems that manage many activity steps in a business process flow. 5. Client applications invoke Corticon decision services by first constructing an XML message or by creating a collection of Java business objects that contain business facts. The client application communicates this information to Corticon Server via SOAP, REST/JSON, JMS, RMI or an in-process call. DEPLOYMENT OPTIONS Corticon decision services conform to SOA. Each decision service: Automates a business decision-making activity Is implemented by a set of business rules Is managed by software developers or business analysts In each enterprise, the application architect needs to determine: How decision services become part of the enterprise architecture Which applications utilize decision services How client applications will invoke the decision services These choices depend on the current and future technical requirements of the enterprise. The available options are summarized in the following table and addressed in detail below: CORTICON DEPLOYMENT OPTION APPROPRIATE FOR MESSAGING OPTIONS Option 1: Remote Server Application Server: Java EE, Spring Applications invoke decision services remotely as web services Service-oriented architectures Distributed architectures Rule modernizing legacy systems with ability to make calls to remote decision services XML SOAP REST/JSON JAX-RPC JAX-WS Option 2: In-process Server: Applications invoke decision services directly via Corticon Server APIs Systems that require the fastest possible performance Systems that require tight coupling of client application with the Corticon Server XML Java objects

3 OPTION 1: REMOTE SERVER This option requires the Corticon Server to be deployed in a multi-threaded Java EE or Spring server. Decision services can be invoked as web services using any of the protocols supported by the application server (e.g., SOAP, REST/JSON, JAX-RPC, JAX-WS). To prepare for web services deployment, the Corticon Deployment Console is used to generate CDDs (Corticon Deployment Descriptor files) and WSDLs/XSDs (service contracts). A CDD is an optional deployment descriptor file which defines a set of decision services to be deployed in Corticon Server. The CDD specifies the locations of rule sets, performance-related parameters and other administrative details that the server needs to effectively deploy the decision services. A WSDL/XSD is effectively a service contract for a decision service. Each decision service has a WSDL/XSD that defines the decision service s expected inputs and outputs. PROGRESS CORTICON SERVER Figure 2 Corticon Server deployed in application server SOAP REST/JSON JAX RPC JAX-WS JAVA APPLICATION SERVER CUSTOM APPLICATIONS OPTION 2: IN-PROCESS SERVER Client applications can invoke decision services and perform administrative functions in-process via the Corticon Server API. This option can deliver high performance but lacks the flexibility and loose coupling that web services can provide. Using Corticon Server as an in-process component, client applications have the option to communicate via XML or Java objects (POJOs). Using Java objects can yield superior performance by eliminating XML translation overhead. PROGRESS CORTICON SERVER Figure 3 Corticon Server deployed in-process XML JAVA APIs REQUEST MESSAGE CUSTOM APPLICATIONS JAVA VIRTUAL MACHINE OBJECTS

4 Out-of-the-box, the Corticon Server is deployed in a Tomcat application server. The Corticon Server also ships with sample Java EE containers (e.g., servlet and session bean), which can be used as-is or customized and deployed in a Java EE or Spring application server. The Corticon Server JARS can also be added to any Java application classpath for in-process use. INVOKING CORTICON DECISION SERVICES Once Corticon decision services have been loaded into Corticon Server, they are ready to be invoked by client applications. The integration approach differs depending upon the selected deployment option. The service requests contain rich payloads that can be carried either via XML or Java objects. The contents of the payloads are analyzed in more depth below. MESSAGING Messaging between a client application and a decision service is conceptually identical across deployment approaches, regardless of the transport protocol used. 1. The client application sends a request message to the Corticon Server, containing a payload of data and targeted at a specific decision service. 2. The Corticon Server invokes the appropriate decision service, which processes its rules against the request payload. 3. When the rules processing is complete, the Corticon Server returns a response message. REQUEST MESSAGE The request message consists of data to be sent to the rules engine. The payload is a collection of application-specific entity instances that are defined in the user s vocabulary. Entity instances may have any number of attributes and may have associations with other entity instances. The collection of entity instances are the facts upon which the rules engine operates. RESPONSE MESSAGE The response message payload includes two discrete sets of data: data payload and posted messages. The data payload contains a copy of the incoming data that has been updated as a consequence of rules firing. Entity instances may be created, updated or removed. In addition, associations between entity instances may be created or removed. In other words, the response message data payload reflects the state of the decision service working memory after all rules have fired. Posted messages contain an audit trail of rules fired by the decision service during the processing of the request, including the sequence of firing. The messages are posted using natural-language rule statements. USING SERVICE CONTRACTS (WSDLS OR XML SCHEMAS) The Corticon request and response messages must adhere to the message structure and data specification as defined in the service contracts. The service contracts can be generated for each Corticon decision service (one service contract per decision service) using the Corticon Deployment Console, as either WSDL or XML schema files (XSDs). Service contracts are helpful

5 integration specifications used by application developers to assemble services into larger applications. WSDL is used when the customer has chosen web services architecture. The WSDL file defines a complete interface description, including: Decision service name Request message structure Response message structure SOAP envelope and binding information XSDs can be used in conjunction with the in-process deployment option. The XSD file is identical to the WSDL file minus the SOAP envelope and binding information. When using XML payloads, the XSD file defines the necessary structure of the XML document (i.e., the XML document can be validated against the XSD using a validating parser). You also have the option to use REST/JSON for executing decision services, in which case you do not have to adhere to a service contract. CORTICON RULE ENGINE Corticon Server employs an engine that has been built specifically to address the needs of business decision automation within the context of a business process. In decision automation, efficient processing of business rules and consistent results are essential to achieve fully automated straight-through-processing ( STP ). Most rule engines analyze rules during execution. This means significant processing is taking place when systems are looking for an answer from the rules system. The traditional pattern matching algorithm to efficiently process and execute rules is Rete. Rete-based inference algorithms are built to address the needs of expert systems in which rules are used for decision support, not decision automation. In those systems a person typically queries a large rule base with a question, and the system converges on zero or more (sometimes conflicting) answers. Because a human interprets the results, logical inconsistencies (such as no answers or conflicting answers) are tolerated and, in fact, architecturally required. In order to achieve STP, decision automation requires consistent answers (guaranteed, conflict-free results) each and every time because typically no human expert is available to be the final arbiter. Furthermore, Rete-based inference engines are designed to allow individual rules to be dynamically added or removed to/from the rule base, as this is a requirement for expert systems. Using a one-activated-rule-at-a-time process to support its dynamic conflict resolution architecture, the Rete algorithm inherently produces a processing bottleneck. While Rete scales well with an increasing number of rules, it degrades exponentially with the increasing complexity of data, resulting in a performance scalability problem well known as the Rete wall. Corticon took a radically different approach for decision automation. Decision services, which are made up of sets of rules, are deployed/undeployed as a unit, so the rule engine does not need to support dynamic tweaking of individual rules. The Corticon engine was designed from the ground-up for decision automation, not decision support. Since Corticon Studio helps the rule author identify and resolve logical conflicts during the rule modeling phase, the Corticon engine s smart compilers automatically generate optimal processing sequences for any decision service. This results in a best-of-both-worlds solution as Corticon delivers blazing performance that scales linearly with both the number of rules and the complexity of the data. Corticon decision service execution is stateless, where all state is maintained in the message payloads. A single incoming message payload seeds the engine working memory after which rules processing commences. Once rules processing ceases, the final state of the engine working memory is returned as an outgoing message payload along with an association rule audit log.

6 ENTERPRISE DATA CONNECTIVITY Corticon supports two distinct modes of operation: 1. Requiring the client program to supply all input data as specified by the service contract 2. Allowing Corticon Server with EDC (enterprise data connectivity) to dynamically access data in an enterprise data source (e.g., relational database) as needed during execution In the first mode of operation, the rule author can decide what information must be passed to the decision service. This is done through a combination of the vocabulary, the rules written using that vocabulary, and the service contract, which specifies the data elements required to process a particular decision service. In the second mode of operation, the client application can supply primary key information in the request message while the rules engine dynamically retrieves (and optionally updates) additional information in an enterprise data source. For example, in a loan pricing application the client may supply the ID of a loan applicant, and, based on rules, the engine may automatically retrieve the applicant s payment history in order to infer new facts such as loan risk. Since the client application supplies only the minimum necessary information, this mode results in looser coupling between the client and the server; rules that refer to enterprise data can evolve freely without affecting the decision service contract or client program logic. Additionally, this mode is very useful when it is difficult to collect in advance all the data necessary to make a decision or if the amount of data required can be substantial in size and difficult to pass into the decision service. In either mode of operation, rules are authored in precisely the same manner. Enterprise mapping specifications are declared in the vocabulary, but those mappings are transparent to the rule author. This key advantage can reduce costs associated with developing and maintaining enterprise rules that require data access. Traditional rule engines typically require technical skills to code database connectivity and the use of SQL to query the database. MULTI-THREADING AND CONCURRENCY Each incoming request message is processed in its own thread of execution: that is, each decision service instance runs in a separate thread. The Corticon Server does not spawn any threads and does not perform thread management; rather, the thread of execution is established and managed by the enclosing container that receives the request (i.e., the application server). In cases where an in-process call is made from the client application, the thread of Corticon execution is the same one in which the client runs. When the Corticon Deployment Console is used, the person responsible for deployment decides how many instances of the same decision service may run concurrently. This is called the decision service pool. Different decision services may have different pool sizes in the same Corticon Server depending on their individual load requirements. CLUSTERING AND LOAD BALANCING In large volume scenarios, enterprises will typically deploy multiple application servers as a cluster. Corticon Server, as a well-behaved, container-managed web service, can leverage this capacity to scale. In addition, a variety of means exist to spread the incoming workload across the multiple application servers. Illustrations of the Corticon Server in active/passive (Figure 4) and active/active (Figure 5) clustering configurations are shown below.

7 PRODUCTION SERVERS Figure 4 Corticon server with active/ passive clustering LOAD BALANCER MESSAGE ONLY IF PRODUCTION SERVERS FAILS CUSTOM APPLICATIONS WARM STANDBY PRODUCTION SERVERS Figure 5 Corticon Server with active/ active clustering LOAD BALANCER MESSAGE CUSTOM APPLICATIONS MAINFRAME INTEROPERABILITY One of the primary reasons for modernizing legacy applications running on the mainframe is the need to provide transparency to the business policies embedded in them. Often, the life of such legacy applications can be extended by externalizing their business logic and automating them as decision services running in Corticon. There are a number of alternatives for invoking an external decision service from a legacy mainframe application including: OPTION 1 (REMOTE): Linux for z/os Java application server (e.g., Tomcat) running on Linux Corticon Server deployed in a Java application server From a COBOL application invoke decision services using SOAP or REST OPTION 2 (REMOTE): WebSphere for z/os Corticon Server deployed in WebSphere From COBOL application invoke decision services using SOAP or REST OPTION 3 (IN-PROCESS): Corticon Server running in a Java Virtual Machine for z/os From an enterprise COBOL application invoke decision services using Corticon Server API

8 Figure 6 shows a schematic of option 2 above: COBOL APPLICATION Figure 6 Legacy mainframe interoperability WEB SPHERE SOAP REST PROGRESS CORTICON STUDIO DEPLOY PROGRESS CORTICON SERVER SUMMARY The Corticon Server supports a variety of enterprise deployment options. Corticon Server s architecture is well-suited for enterprise environments, leveraging the inherent scalability, availability, and manageability of enterprise-grade application servers. Corticon-based applications can easily adjust to changes in the business landscape and can help organizations achieve the benefits of a service-oriented architecture. PROGRESS SOFTWARE Progress Software Corporation (NASDAQ: PRGS) is a global software company that simplifies the development, deployment and management of business applications onpremise or in the cloud, on any platform or device, to any data source, with enhanced performance, minimal IT complexity and low total cost of ownership. WORLDWIDE HEADQUARTERS Progress Software Corporation, 14 Oak Park, Bedford, MA 01730 USA Tel: +1 781 280-4000 Fax: +1 781 280-4095 On the Web at: Find us on facebook.com/progresssw twitter.com/progresssw youtube.com/progresssw For regional international office locations and contact information, please go to /worldwide Progress and Corticon are trademarks or registered trademarks of Progress Software Corporation or one of its affiliates or subsidiaries in the U.S. and other countries. Any other marks contained herein may be trademarks of their respective owners. Specifications subject to change without notice. 2011-2014 Progress Software Corporation. All rights reserved. Rev. 07/14 140623-0066