+ All Categories
Home > Documents > ISM Notes Part II

ISM Notes Part II

Date post: 04-Apr-2018
Category:
Upload: geet-goel
View: 219 times
Download: 0 times
Share this document with a friend

of 20

Transcript
  • 7/30/2019 ISM Notes Part II

    1/20

  • 7/30/2019 ISM Notes Part II

    2/20

    Key issues in Supply Chain Management-

    ISSUE CONSIDERATIONS

    Network Planning Warehouse locations and capacities

    Plant locations and production levels Transportation flows between facilities to minimize cost a

    time

    Inventory Control How should inventory be managed? Why does inventory fluctuate and what strategies minimiz

    this?

    Supply Contracts Impact of volume discount and revenue sharing

    Pricing strategies to reduce order-shipment variability

    Distribution Strategies Selection of distribution strategies (e.g., direct ship vs. cro

    docking)

    How many cross-dock points are needed?

    Cost/Benefits of different strategiesIntegration and Strategic Partnering How can integration with partners be achieved?

    What level of integration is best? What information and processes can be shared?

    What partnerships should be implemented and in which

    situations?

    Outsourcing & Procurement

    Strategies

    What are our core supply chain capabilities and which are

    not? Does our product design mandate different outsourcing

    approaches?

    Risk management

    Product Design How are inventory holding and transportation costs affecteby product design?

    How does product design enable mass customization?

    Supply Chain Management Strategies

    STRATEGY WHEN TO CHOOSE BENEFITS

    Make to Stock standardized products,relatively predictable demand

    Low manufacturing costs;meet customer demands

    quickly

    Make to Order customized products, many

    variations

    Customization; reduced

    inventory; improved service

    levels

    Configure to Order many variations on finished

    product; infrequent demand

    Low inventory levels; wide

    range of product offerings;simplified planning

    Engineer to Order complex products, unique

    customer specifications

    Enables response to specific

    customer requirements

  • 7/30/2019 ISM Notes Part II

    3/20

    Supply chain management must address the following problems:

    Distribution Network Configuration: number, location and network missions of suppliers,

    production facilities, distribution centers, warehouses, cross-docks and customers.

    Distribution Strategy: questions of operating control (centralized, decentralized or shared);

    delivery scheme, e.g., direct shipment, pool point shipping, cross docking, direct storedelivery (DSD), closed loop shipping; mode of transportation, e.g., motor carrier, includingtruckload, Less than truckload (LTL), parcel; railroad; intermodal transport, including trailer

    on flatcar (TOFC) and container on flatcar (COFC); ocean freight; airfreight; replenishment

    strategy (e.g., pull, push or hybrid); and transportation control (e.g., owner-operated, privatecarrier, common carrier, contract carrier, orthird-party logistics (3PL)).

    Trade-Offs in Logistical Activities: The above activities must be well coordinated in order

    to achieve the lowest total logistics cost. Trade-offs may increase the total cost if only one of

    the activities is optimized. For example, full truckload (FTL) rates are more economical on acost per pallet basis than LTL shipments. If, however, a full truckload of a product is ordered

    to reduce transportation costs, there will be an increase in inventory holding costs which may

    increase total logistics costs. It is therefore imperative to take a systems approach whenplanning logistical activities. These trade-offs are key to developing the most efficient and

    effective Logistics and SCM strategy.

    Information: Integration of processes through the supply chain to share valuableinformation, including demand signals, forecasts, inventory, transportation, potential

    collaboration, etc.

    Inventory Management: Quantity and location of inventory, including raw materials, work-

    in-process (WIP) and finished goods. Cash-Flow: Arranging the payment terms and methodologies for exchanging funds across

    entities within the supply chain.

    Supply chain execution means managing and coordinating the movement of materials, informationand funds across the supply chain. The flow is bi-directional.

    Supply chain management Activities

    Supply chain management is a cross-function approach including managing the movement of raw

    materials into an organization, certain aspects of the internal processing of materials into finished

    goods, and the movement of finished goods out of the organization and toward the end-consumer.As organizations strive to focus on core competencies and becoming more flexible, they reduce their

    ownership of raw materials sources and distribution channels. These functions are increasingly being

    outsourced to other entities that can perform the activities better or more cost effectively. The effect

    is to increase the number of organizations involved in satisfying customer demand, while reducingmanagement control of daily logistics operations. Less control and more supply chain partners led to

    the creation of supply chain management concepts. The purpose of supply chain management is toimprove trust and collaboration among supply chain partners, thus improving inventory visibility

    and the velocity of inventory movement.

    Several models have been proposed for understanding the activities required to manage material

    movements across organizational and functional boundaries. SCOR is a supply chain managementmodel promoted by the Supply Chain Council. Another model is the SCM Model proposed by the

    Global Supply Chain Forum (GSCF). Supply chain activities can be grouped into strategic, tactical,

    and operational levels

    http://en.wikipedia.org/wiki/Direct_shipmenthttp://en.wikipedia.org/wiki/Cross_dockinghttp://en.wikipedia.org/wiki/Motor_carrierhttp://en.wikipedia.org/wiki/Less_than_truckloadhttp://en.wikipedia.org/wiki/Railroadhttp://en.wikipedia.org/wiki/Private_carrierhttp://en.wikipedia.org/wiki/Private_carrierhttp://en.wikipedia.org/wiki/Common_carrierhttp://en.wikipedia.org/wiki/Third-party_logisticshttp://en.wikipedia.org/wiki/Direct_shipmenthttp://en.wikipedia.org/wiki/Cross_dockinghttp://en.wikipedia.org/wiki/Motor_carrierhttp://en.wikipedia.org/wiki/Less_than_truckloadhttp://en.wikipedia.org/wiki/Railroadhttp://en.wikipedia.org/wiki/Private_carrierhttp://en.wikipedia.org/wiki/Private_carrierhttp://en.wikipedia.org/wiki/Common_carrierhttp://en.wikipedia.org/wiki/Third-party_logistics
  • 7/30/2019 ISM Notes Part II

    4/20

    Strategic level

    Strategic network optimization, including the number, location, and size of warehousing,

    distribution centers, and facilities.

    Strategic partnerships with suppliers, distributors, and customers, creating communication

    channels for critical information and operational improvements such as cross docking, direct

    shipping, and third-party logistics.

    Product life cycle management, so that new and existing products can be optimally integratedinto the supply chain and capacity management activities.

    Segmentation of products and customers to guide alignment of corporate objectives withmanufacturing and distribution strategy.

    Information technology chain operations.

    Where-to-make and make-buy decisions.

    Aligning overall organizational strategy with supply strategy. It is for long term and needs resource commitment.

    Tactical level

    Sourcing contracts and other purchasing decisions. Production decisions, including contracting, scheduling, and planning process definition.

    Inventory decisions, including quantity, location, and quality of inventory.

    Transportation strategy, including frequency, routes, and contracting.

    Benchmarking of all operations against competitors and implementation of best practicesthroughout the enterprise.

    Milestone payments.

    Focus on customer demand and Habits.

    Operational level

    Daily production and distribution planning, including all nodes in the supply chain. Production scheduling for each manufacturing facility in the supply chain (minute by

    minute).

    Demand planning and forecasting, coordinating the demand forecast of all customers and

    sharing the forecast with all suppliers.

    Sourcing planning, including current inventory and forecast demand, in collaboration with allsuppliers.

    Inbound operations, including transportation from suppliers and receiving inventory.

    Production operations, including the consumption of materials and flow of finished goods.

    Outbound operations, including all fulfillment activities, warehousing and transportation tocustomers.

    Order promising, accounting for all constraints in the supply chain, including all suppliers,manufacturing facilities, distribution centers, and other customers.

    From production level to supply level accounting all transit damage cases & arrange to

    settlement at customer level by maintaining company loss through insurance company.

    Managing non-moving, short-dated inventory and avoiding more products to go short-

    dated.

    http://en.wikipedia.org/wiki/Distribution_centerhttp://en.wikipedia.org/wiki/Strategic_partnershiphttp://en.wikipedia.org/wiki/Cross_dockinghttp://en.wikipedia.org/wiki/Third-party_logisticshttp://en.wikipedia.org/wiki/Product_life_cycle_managementhttp://en.wikipedia.org/wiki/Information_technologyhttp://en.wikipedia.org/wiki/Benchmarkinghttp://en.wikipedia.org/wiki/Best_practicehttp://en.wikipedia.org/wiki/Distribution_centerhttp://en.wikipedia.org/wiki/Strategic_partnershiphttp://en.wikipedia.org/wiki/Cross_dockinghttp://en.wikipedia.org/wiki/Third-party_logisticshttp://en.wikipedia.org/wiki/Product_life_cycle_managementhttp://en.wikipedia.org/wiki/Information_technologyhttp://en.wikipedia.org/wiki/Benchmarkinghttp://en.wikipedia.org/wiki/Best_practice
  • 7/30/2019 ISM Notes Part II

    5/20

    Customer Relationship Management (CRM)

    Customer Relationship Management (CRM) can be widely defined as company activities related todeveloping and retaining customers. It is a blend of internal business processes: sales, marketing and

    customer support with technology and data capturing techniques. Customer Relationship

    Management is all about building long-term business relationships with customers.

    CRM is an alignment of strategy, processes and technology to manage customers and all customer-

    facing departments & partners. Any CRM initiative is and has the potential of providing strategic

    advantages to the organization, if handled right.

    It is a process or methodology used to learn more about customers' needs and behaviors in order to

    develop stronger relationships with them. There are many technological components to CRM, butthinking about CRM in primarily technological terms is a mistake. The more useful way to think

    about CRM is as a process that will help bring together lots of pieces of information about

    customers, sales, marketing effectiveness, responsiveness and market trends.CRM helps businesses use technology and human resources to gain insight into the behavior of

    customers and the value of those customers.

    Advantages Of CRM

    1. Using CRM, a business can:

    2. Provide better customer service

    3. Increase customer revenues

    4. Discover new customers

    5. Cross sell/Up Sell products more effectively

    6. Help sales staff close deals faster

    7. Make call centers more efficient

    8. Simplify marketing and sales processes

    Other advantages are-

    CRM solutions help companies boost their business efficiency, thereby increasing profit

    and revenue generation capabilities. Let us take a quick look at some of the measurable

    benefits that your organization can gain by implementing a CRM solution.

    Increase Customer Lifecycle Value

    In most businesses, the cost of acquisition of customers is high. To make profits, it is

    important to keep the customer longer and sell him more products (cross sell, up sell,

    etc) to him, during his lifecycle. Customer stay, if they are provided with value, quality

  • 7/30/2019 ISM Notes Part II

    6/20

    service and continuity. CRM solutions enable you to do that.

    Execution Control

    Once the business strategy is put into motion, the management needs feedback and

    reports to judge how the business is performing. CRM solutions provide management

    with control and a scientific way to identify and resolve issues. The benefits include aclearer visibility of the sales pipeline, accurate forecasts and more.

    Customer Lifecycle Management

    To keep the customers happy, you need to know them better. At the minimum, you need

    a centralize customer database, that captures most of the information from your entire

    customer facing departments and partners. Integrated CRM solutions, like CRMnextenable you to manage customer information, throughout all stages of their life cycle,

    from contact to contract to customer service.

    Strategic Consistency

    Because CRM offers business and technological alignment, it enables companies toachieve strategic company goals more effectively, like enhanced sales realization, higher

    customer satisfaction, better brand management and more. Additionally, the alignment

    results in a more consistent customer communication creating a feeling of continuity.

    Business Intelligence

    Due to the valuable business insights that CRM provides, it becomes easier to identifythe bottlenecks, their causes and the remedial measures that need to be taken. For

    example, CRMnext provides real-time business focus dashboards with extensive drill

    down capabilities that provide the decision makers with the depth of information

    required to identify the causes and spot trends.

    Definition 1-Data Warehouse

    A data warehouse is a collection and summarization of information from multiple databases anddatabase tables. The primary purpose of a data warehouse is not data storage, but the

    collection of information for decision-making. Typically, a data warehouse extracts updated

    information from operational databases on a regular basis (nightly, hourly, etc.). This forms a

    snapshot of collected data that can be organized into a logical structure based on youranalytical needs.

    Data warehouses allow you to express your information needs logically, without beingconstrained to database fields and records. Using the correct data mining tools, it is possible to

    display information from a data warehouse in ways that are not possible using SQL or other

    basic query languages. Unlike a relational database, a data warehouse can present informationin multidimensional format. This representation is called a hypercube, and contains layers of

    rows and columns. Using this model a company could, for instance, track sales of multiple

    products in multiple regions over a given period of time, all in the same view.

  • 7/30/2019 ISM Notes Part II

    7/20

    A data warehouse can contain extremely large amounts of information, and many users will

    only need to access a portion of this. Information in a data warehouse can be organized into

    data marts, which are subsets of data with a specific focus. Data marts can provide an analystwith a more efficient set of working data relevant to, for instance, a specific business process

    or unit of the company

    Definition 2- Data Warehouse?A data warehouse is a relational database that is designed for query and analysis rather than fortransaction processing. It usually contains historical data derived from transaction data, but it can

    include data from other sources. It separates analysis workload from transaction workload and

    enables an organization to consolidate data from several sources.

    In addition to a relational database, a data warehouse environment includes an extraction,transportation, transformation, and loading (ETL) solution, an online analytical processing (OLAP)

    Supplier Database

    Data warehouse

    Customer DatabaseSales Database

    Data Mart

    Data Mart

  • 7/30/2019 ISM Notes Part II

    8/20

    engine, client analysis tools, and other applications that manage the process of gathering data and

    delivering it to business users

    Different types of data warehouse :

    Subject Oriented

    Integrated

    Nonvolatile Time Variant

    Subject Oriented

    Data warehouses are designed to help you analyze data. For example, to learn more about yourcompany's sales data, you can build a warehouse that concentrates on sales. Using this warehouse,

    you can answer questions like "Who was our best customer for this item last year?" This ability to

    define a data warehouse by subject matter, sales in this case, makes the data warehouse subjectoriented.

    Integrated

    Integration is closely related to subject orientation. Data warehouses must put data from disparate

    sources into a consistent format. They must resolve such problems as naming conflicts andinconsistencies among units of measure. When they achieve this, they are said to be integrated.

    Nonvolatile

    Nonvolatile means that, once entered into the warehouse, data should not change. This is logical

    because the purpose of a warehouse is to enable you to analyze what has occurred.

    Time Variant

    In order to discover trends in business, analysts need large amounts of data. This is very much in

    contrast to online transaction processing (OLTP) systems, where performance requirements

    demand that historical data be moved to an archive. A data warehouse's focus on change over time iswhat is meant by the term time variant

    Data Warehouse Architectures

    Data warehouses and their architectures vary depending upon the specifics of an organization's

    situation. Three common architectures are:

    Data Warehouse Architecture (Basic)

    Data Warehouse Architecture (with a Staging Area)

    Data Warehouse Architecture (with a Staging Area and Data Marts)

    http://docs.oracle.com/cd/B10501_01/server.920/a96520/concept.htm#49840http://docs.oracle.com/cd/B10501_01/server.920/a96520/concept.htm#49842http://docs.oracle.com/cd/B10501_01/server.920/a96520/concept.htm#49871http://docs.oracle.com/cd/B10501_01/server.920/a96520/concept.htm#49853http://docs.oracle.com/cd/B10501_01/server.920/a96520/glossary.htm#432248http://docs.oracle.com/cd/B10501_01/server.920/a96520/glossary.htm#432248http://docs.oracle.com/cd/B10501_01/server.920/a96520/concept.htm#51090http://docs.oracle.com/cd/B10501_01/server.920/a96520/concept.htm#50822http://docs.oracle.com/cd/B10501_01/server.920/a96520/concept.htm#51078http://docs.oracle.com/cd/B10501_01/server.920/a96520/concept.htm#49840http://docs.oracle.com/cd/B10501_01/server.920/a96520/concept.htm#49842http://docs.oracle.com/cd/B10501_01/server.920/a96520/concept.htm#49871http://docs.oracle.com/cd/B10501_01/server.920/a96520/concept.htm#49853http://docs.oracle.com/cd/B10501_01/server.920/a96520/glossary.htm#432248http://docs.oracle.com/cd/B10501_01/server.920/a96520/concept.htm#51090http://docs.oracle.com/cd/B10501_01/server.920/a96520/concept.htm#50822http://docs.oracle.com/cd/B10501_01/server.920/a96520/concept.htm#51078
  • 7/30/2019 ISM Notes Part II

    9/20

    Data Warehouse Architecture (Basic)

    Figure shows a simple architecture for a data warehouse. End users directly access data derived

    from several source systems through the data warehouse.

    Figure 1- Architecture of a Data Warehouse

    In Figure 1 the metadata and raw data of a traditional OLTP system is present, as is an additional

    type of data, summary data. Summaries are very valuable in data warehouses because they pre-compute long operations in advance. For example, a typical data warehouse query is to retrieve

    something like August sales.

    Data Warehouse Architecture (with a Staging Area)

    In above Figure1 , you need to clean and process your operational data before putting it into thewarehouse. You can do this programmatically, although most data warehouses use a staging area-instead. A staging area simplifies building summaries and general warehouse management. Figure 2

    illustrates this typical architecture.

    http://docs.oracle.com/cd/B10501_01/server.920/a96520/glossary.htm#433213http://docs.oracle.com/cd/B10501_01/server.920/a96520/glossary.htm#433213
  • 7/30/2019 ISM Notes Part II

    10/20

    Figure 2- Architecture of a Data Warehouse with a Staging Area

    Data Warehouse Architecture (with a Staging Area and Data Marts)

    Although the architecture in Figure 2 is quite common, you may want to customize your

    warehouse's architecture for different groups within your organization. You can do this by adding

    datamarts, which are systems designed for a particular line of business. Figure 3 illustrates anexample where purchasing, sales, and inventories are separated. In this example, a financial analyst

    might want to analyze historical data for purchases and sales.

    Figure 3- Architecture of a Data Warehouse with a Staging Area and Data Marts

  • 7/30/2019 ISM Notes Part II

    11/20

    OLTP (On-lineTransaction Processing) is characterized by a large number of short on-line

    transactions (INSERT, UPDATE, DELETE). The main emphasis for OLTP systems is put on veryfast query processing, maintaining data integrity in multi-access environments and an effectiveness

    measured by number of transactions per second. In OLTP database there is detailed and current

    data, and schema used to store transactional databases is the entity model (usually 3NF).

    - OLAP (On-line Analytical Processing) is characterized by relatively low volume of transactions.

    Queries are often very complex and involve aggregations. For OLAP systems a response time is an

    effectiveness measure. OLAP applications are widely used by Data Mining techniques. In OLAPdatabase there is aggregated, historical data, stored in multi-dimensional schemas (usually star

    schema).

    The following table summarizes the major differences between OLTP and OLAP system design.

    OLTP System

    Online Transaction

    Processing

    (Operational System)

    OLAP System

    Online Analytical Processing

    (Data Warehouse)

    Source of dataOperational data; OLTPs are the original

    source of the data.Consolidation data; OLAP data comes from

    the various OLTP Databases

    Purpose of dataTo control and run fundamental business

    tasks

    To help with planning, problem solving, and

    decision support

    What the dataReveals a snapshot of ongoing business

    processes

    Multi-dimensional views of various kinds of

    business activitiesInserts and

    Updates

    Short and fast inserts and updates initiated

    by end users

    Periodic long-running batch jobs refresh the

    data

    QueriesRelatively standardized and simple

    queries Returning relatively few recordsOften complex queries involving

    aggregations

    Processing

    SpeedTypically very fast

    Depends on the amount of data involved;batch data refreshes and complex queries

    may take many hours; query speed can be

    improved by creating indexes

    SpaceRequirements

    Can be relatively small if historical data isarchived

    Larger due to the existence of aggregationstructures and history data; requires more

    indexes than OLTP

    DatabaseDesign

    Highly normalized with many tablesTypically de-normalized with fewer tables;

    use of star and/or snowflake schemas

    Backup andRecovery

    Backup religiously; operational data is

    critical to run the business, data loss islikely to entail significant monetary loss

    and legal liability

    Instead of regular backups, someenvironments may consider simply reloading

    the OLTP data as a recovery method

    http://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.htmlhttp://datawarehouse4u.info/OLTP-vs-OLAP.html
  • 7/30/2019 ISM Notes Part II

    12/20

    Difference between data Warehouse and database

    The primary difference betwen you application database and a data warehouse is that while theformer is designed (and optimized) to record , the latter has to be designed (and optimized) to

    respond to analysis questions that are critical for your business.

    Application databases are OLTP (On-Line Transaction Processing) systems where every transactionhas to be recorded, and super-fast at that. Consider the scenario where a bank ATM has disbursed

    cash to a customer but was unable to record this event in the bank records. If this started happening

    frequently, the bank wouldn't stay in business for too long. So the banking system is designed tomake sure that every trasaction gets recorded within the time you stand before the ATM machine.

    This system is write-optimized, and you shouldn't crib if your analysis query (read operation) takes a

    lot of time on such a system.

    A Data Warehouse (DW) on the other end, is a database that is designed for facilitating querying and

    analysis. Often designed as OLAP (On-Line Analytical Processing) systems, these databases containread-only data that can be queried and analysed far more efficiently as compared to your regular

    OLTP application databases. In this sense an OLAP system is designed to be read-optimized.

    Separation from your application database also ensures that your business intelligence solution isscalable (your bank and ATMs don't go down just because the CFO asked for a report), better

    documented and managed (god help the novice who is given the application database diagrams and

    asked to locate the needle of data in the proverbial haystack of table proliferation), and can answerquestions far more efficietly and frequently.

    Creation of a DW leads to a direct increase in quality of analyses as the table structures are simpler

    (you keep only the needed information in simpler tables), standardized (well-documented tablestructures), and often denormalized (to reduce the linkages between tables and the corresponding

    complexity of queries). A DW drastically reduces the 'cost-per-analysis' and thus permits more

    analysis per FTE. Having a well-designed DW is the foundation successful BI/Analytics initiativesare built upon.

    Data Mining

    Data mining (the analysis step of the "Knowledge Discovery in Databases" process, or KDD),a

    field at the intersection of computer science and statistics,is the process that attempts to discoverpatterns in large data sets. It utilizes methods at the intersection of artificial intelligence, machine

    learning, statistics, and database systems. The overall goal of the data mining process is to extractinformation from a data set and transform it into an understandable structure for further use. Aside

    from the raw analysis step, it involves database and data management aspects, data preprocessing,

    model and inference considerations, interestingness metrics, complexity considerations, post-processing of discovered structures, visualization, and online updating

    The term is a buzzword, and is frequently misused to mean any form of large-scale data or

    information processing (collection, extraction, warehousing, analysis, and statistics) but is also

    http://en.wikipedia.org/wiki/Computer_sciencehttp://en.wikipedia.org/wiki/Statisticshttp://en.wikipedia.org/wiki/Data_sethttp://en.wikipedia.org/wiki/Artificial_intelligencehttp://en.wikipedia.org/wiki/Machine_learninghttp://en.wikipedia.org/wiki/Machine_learninghttp://en.wikipedia.org/wiki/Statisticshttp://en.wikipedia.org/wiki/Database_systemhttp://en.wikipedia.org/wiki/Data_managementhttp://en.wikipedia.org/wiki/Data_Pre-processinghttp://en.wikipedia.org/wiki/Statistical_modelhttp://en.wikipedia.org/wiki/Statistical_inferencehttp://en.wikipedia.org/wiki/Computational_complexity_theoryhttp://en.wikipedia.org/wiki/Data_visualizationhttp://en.wikipedia.org/wiki/Online_algorithmhttp://en.wikipedia.org/wiki/Buzzwordhttp://en.wikipedia.org/wiki/Information_extractionhttp://en.wikipedia.org/wiki/Data_warehousehttp://en.wikipedia.org/wiki/Data_analysishttp://en.wikipedia.org/wiki/Computer_sciencehttp://en.wikipedia.org/wiki/Statisticshttp://en.wikipedia.org/wiki/Data_sethttp://en.wikipedia.org/wiki/Artificial_intelligencehttp://en.wikipedia.org/wiki/Machine_learninghttp://en.wikipedia.org/wiki/Machine_learninghttp://en.wikipedia.org/wiki/Statisticshttp://en.wikipedia.org/wiki/Database_systemhttp://en.wikipedia.org/wiki/Data_managementhttp://en.wikipedia.org/wiki/Data_Pre-processinghttp://en.wikipedia.org/wiki/Statistical_modelhttp://en.wikipedia.org/wiki/Statistical_inferencehttp://en.wikipedia.org/wiki/Computational_complexity_theoryhttp://en.wikipedia.org/wiki/Data_visualizationhttp://en.wikipedia.org/wiki/Online_algorithmhttp://en.wikipedia.org/wiki/Buzzwordhttp://en.wikipedia.org/wiki/Information_extractionhttp://en.wikipedia.org/wiki/Data_warehousehttp://en.wikipedia.org/wiki/Data_analysis
  • 7/30/2019 ISM Notes Part II

    13/20

    generalized to any kind of computer decision support system, including artificial intelligence,

    machine learning, andbusiness intelligence. In the proper use of the word, the key term is discovery,

    commonly defined as "detecting something new". Even the popular book "Data mining: Practicalmachine learning tools and techniques with Java" (which covers mostly machine learning material)

    was originally to be named just "Practical machine learning", and the term "data mining" was only

    added for marketing reasons. Often the more general terms "(large scale) data analysis", or"analytics" or when referring to actual methods, artificial intelligence and machine learning are

    more appropriate.

    The actual data mining task is the automatic or semi-automatic analysis of large quantities of data to

    extract previously unknown interesting patterns such as groups of data records (cluster analysis),unusual records (anomaly detection) and dependencies (association rule mining). This usually

    involves using database techniques such as spatial indexes. These patterns can then be seen as a kind

    of summary of the input data, and may be used in further analysis or, for example, in machinelearning and predictive analytics. For example, the data mining step might identify multiple groups

    in the data, which can then be used to obtain more accurate prediction results by a decision support

    system. Neither the data collection, data preparation, nor result interpretation and reporting are partof the data mining step, but do belong to the overall KDD process as additional steps.

    Knowledge Discovery in Databases (KDD)

    Knowledge Discovery in Databases (KDD), refers to the nontrivial extraction of implicit, previously

    unknown and potentially useful information from data in databases. While data mining andknowledge discovery in databases (or KDD) are frequently treated as synonyms, data mining is

    actually part of the knowledge discovery process.

    The Knowledge Discovery in Databases (KDD) process is commonly defined with the stages:

    (1) Selection

    (2) Pre-processing(3) Transformation

    (4)Data Mining

    (5) Interpretation/Evaluation.

    It exists, however, in many variations on this theme, such as the Cross Industry Standard Process forData Mining (CRISP-DM) which defines six phases:

    (1) Business Understanding

    (2) Data Understanding(3) Data Preparation

    (4) Modeling

    (5) Evaluation

    (6) Deployment

    or a simplified process such as (1) pre-processing, (2) data mining, and (3) results validation.

    http://en.wikipedia.org/wiki/Decision_support_systemhttp://en.wikipedia.org/wiki/Artificial_intelligencehttp://en.wikipedia.org/wiki/Machine_learninghttp://en.wikipedia.org/wiki/Business_intelligencehttp://en.wikipedia.org/wiki/Discovery_(observation)http://en.wikipedia.org/wiki/Machine_learninghttp://en.wikipedia.org/wiki/Data_analysishttp://en.wikipedia.org/wiki/Analyticshttp://en.wikipedia.org/wiki/Artificial_intelligencehttp://en.wikipedia.org/wiki/Machine_learninghttp://en.wikipedia.org/wiki/Cluster_analysishttp://en.wikipedia.org/wiki/Anomaly_detectionhttp://en.wikipedia.org/wiki/Association_rule_mininghttp://en.wikipedia.org/wiki/Spatial_indexhttp://en.wikipedia.org/wiki/Machine_learninghttp://en.wikipedia.org/wiki/Machine_learninghttp://en.wikipedia.org/wiki/Predictive_analyticshttp://en.wikipedia.org/wiki/Decision_support_systemhttp://en.wikipedia.org/wiki/Decision_support_systemhttp://en.wikipedia.org/wiki/CRISP-DMhttp://en.wikipedia.org/wiki/CRISP-DMhttp://en.wikipedia.org/wiki/Decision_support_systemhttp://en.wikipedia.org/wiki/Artificial_intelligencehttp://en.wikipedia.org/wiki/Machine_learninghttp://en.wikipedia.org/wiki/Business_intelligencehttp://en.wikipedia.org/wiki/Discovery_(observation)http://en.wikipedia.org/wiki/Machine_learninghttp://en.wikipedia.org/wiki/Data_analysishttp://en.wikipedia.org/wiki/Analyticshttp://en.wikipedia.org/wiki/Artificial_intelligencehttp://en.wikipedia.org/wiki/Machine_learninghttp://en.wikipedia.org/wiki/Cluster_analysishttp://en.wikipedia.org/wiki/Anomaly_detectionhttp://en.wikipedia.org/wiki/Association_rule_mininghttp://en.wikipedia.org/wiki/Spatial_indexhttp://en.wikipedia.org/wiki/Machine_learninghttp://en.wikipedia.org/wiki/Machine_learninghttp://en.wikipedia.org/wiki/Predictive_analyticshttp://en.wikipedia.org/wiki/Decision_support_systemhttp://en.wikipedia.org/wiki/Decision_support_systemhttp://en.wikipedia.org/wiki/CRISP-DMhttp://en.wikipedia.org/wiki/CRISP-DM
  • 7/30/2019 ISM Notes Part II

    14/20

    Explain the KDD process.

    Knowledge discovery as a process is depicted and consists of an iterative sequence of the followingsteps:

    1. Data cleaning: to remove noise and inconsistent data

    2. Data integration: where multiple data sources may be combined

    3. Data selection: where data relevant to the analysis task are retrieved from the database

  • 7/30/2019 ISM Notes Part II

    15/20

    4. Data transformation: where data are transformed or consolidated into forms appropriate for

    mining by performing summary or aggregation operations.

    5. Data mining: an essential process where intelligent methods are applied in order to extractdata pattern.

    6. Pattern evaluation to identify the truly interesting patterns representing knowledge based on

    some interestingness measures;7. Knowledge presentation where visualization and knowledge representation techniques are

    used to present the mined knowledge to the user.

    Steps 1 to 4 are different forms of data preprocessing, where the data are prepared for mining. Thedata mining step may interact with the user or a knowledge base.

    The interesting patterns are presented to the user and may be stored as new knowledge in the

    knowledge base. Data mining is only one step in the entire process but an essential one because it

    uncovers hidden patterns for evaluation.

    Therefore, data mining is a step in the knowledge discovery process

    Data Mining

    Data mining is the process of extracting information from large sources of data, such as a corporatedata warehouse, and extrapolating relationships and trends within that data. It is not possible to use

    standard query tools, such as SQL, to perform these operations.

    There are three main categories of data mining tools: query-and-reporting tools, intelligent agents,and multidimensional analysis tools.

    Query-and-reporting tools offer functionality similar to query and report generators for standard

    databases. These tools are easy to use, but their scope is limited to that of a relational database, andthey do not take full advantage of the potential of a data warehouse.

    The term 'intelligent agents' encompasses a variety of artificial intelligence tools which haverecently emerged into the field of data manipulation. Two of these tools are neural networks and

    fuzzy logic. An intelligent agent can sift through the contents of a database, finding unsuspected

    trends and relationships between data.

    Multidimensional analysis tools allow a user to interpret multidimensional data (i.e., a hypercube

    data set) from different perspectives. For example, if a set of data includes products sold in various

    regions over time, multidimensional analysis allows you to view the data in different ways. Forinstance, you could display all sales in all regions for a given time, or all sales over time in a given

    region

  • 7/30/2019 ISM Notes Part II

    16/20

    Data mining involves six common classes of tasks:[1]

    Anomaly detection (Outlier/change/deviation detection) The identification of unusual data

    records, that might be interesting or data errors and require further investigation.

    Association rule learning (Dependency modeling) Searches for relationships between

    variables. For example a supermarket might gather data on customer purchasing habits.Using association rule learning, the supermarket can determine which products are frequently

    bought together and use this information for marketing purposes. This is sometimes referred

    to as market basket analysis.

    Clustering is the task of discovering groups and structures in the data that are in some wayor another "similar", without using known structures in the data.

    Classification is the task of generalizing known structure to apply to new data. For

    example, an e-mail program might attempt to classify an e-mail as "legitimate" or as "spam".

    Regression Attempts to find a function which models the data with the least error.

    Summarization providing a more compact representation of the data set, includingvisualization and report generation.

    Data Warehouse

    Engine

    Query And reporting ToolsMultidimensional Analysis

    Tools

    Intelligent Agent

    Data warehouse

    http://en.wikipedia.org/wiki/Data_mining#cite_note-Fayyad-0http://en.wikipedia.org/wiki/Anomaly_detectionhttp://en.wikipedia.org/wiki/Association_rule_learninghttp://en.wikipedia.org/wiki/Cluster_analysishttp://en.wikipedia.org/wiki/Statistical_classificationhttp://en.wikipedia.org/wiki/Regression_analysishttp://en.wikipedia.org/wiki/Automatic_summarizationhttp://en.wikipedia.org/wiki/Data_mining#cite_note-Fayyad-0http://en.wikipedia.org/wiki/Anomaly_detectionhttp://en.wikipedia.org/wiki/Association_rule_learninghttp://en.wikipedia.org/wiki/Cluster_analysishttp://en.wikipedia.org/wiki/Statistical_classificationhttp://en.wikipedia.org/wiki/Regression_analysishttp://en.wikipedia.org/wiki/Automatic_summarization
  • 7/30/2019 ISM Notes Part II

    17/20

    Architecture of a typical Data Mining System-

    The architecture of a typical data mining system may have the following major components :

    Database, data warehouse, WorldWideWeb, or other information repository:

    This is one or a set of databases, data warehouses, spreadsheets, or other kinds

    of information repositories. Data cleaning and data integration techniques may be performed on the data.

    Database or data warehouse server:

    The database or data warehouse server is responsible for fetching the relevant

    data, based on the users data mining request.

    Knowledge base:

    This is the domain knowledge that is used to guide the search or evaluate the

    interestingness of resulting patterns. Such knowledge can include concept hierarchies, used to organize attributesor attribute values into different levels of abstraction.

    Knowledge such as user beliefs, which can be used to assess a patterns

    interestingness based on its unexpectedness, may also be included.

    Other examples of domain knowledge are additional interestingness

    constraints or thresholds, and metadata (e.g., describing data from multiple

    heterogeneous sources).

    Data mining engine:

    This is essential to the data mining system and ideally consists of a set offunctional modules for tasks such as characterization, association and correlation

    analysis, classification, prediction, cluster analysis, outlier analysis, and evolution

    analysis.

    Pattern evaluation module:

    This component typically employs interestingness measures and interacts with

    the data mining modules so as to focus the search toward interesting patterns. It may

    use interestingness thresholds to filter out discovered patterns.

    Alternatively, the pattern evaluation module may be integrated with the

    mining module, depending on the implementation of the data mining method used. For efficient data mining, it is highly recommended to push the evaluation of

    pattern interestingness as deep as possible into the mining process so as to confine thesearch to only the interesting patterns.

    User interface:

    This module communicates between users and the data mining system,

    allowing the user to interact with the system by specifying a data mining query or

  • 7/30/2019 ISM Notes Part II

    18/20

    task, providing information to help focus the search, and performing exploratory data

    mining based on the intermediate data mining results.

    Also, it allows the user to browse database and data warehouse schemas ordata structures, evaluate mined patterns, and visualize the patterns in different forms.

    Figure-Typical data Mining Architecture

    Classification of Data mining System

    There are many data mining systems available or being developed. Some are specialized systemsdedicated to a given data source or are confined to limited data mining functionalities, other aremore versatile and comprehensive. Data mining systems can be categorized according to various

    criteria among other classification are the following:

    Classification according to the type of data source mined: this classification categorizesdata mining systems according to the type of data handled such as spatial data, multimedia

    data, time-series data, text data, World Wide Web, etc.

  • 7/30/2019 ISM Notes Part II

    19/20

    Classification according to the data model drawn on: this classification categorizes data

    mining systems based on the data model involved such as relational database, object-oriented

    database, data warehouse, transactional, etc.

    Classification according to the king of knowledge discovered: this classification

    categorizes data mining systems based on the kind of knowledge discovered or data mining

    functionalities, such as characterization, discrimination, association, classification,clustering, etc. Some systems tend to be comprehensive systems offering several data mining

    functionalities together. Classification according to mining techniques used: Data mining systems employ and

    provide different techniques. This classification categorizes data mining systems according

    to the data analysis approach used such as machine learning, neural networks, genetic

    algorithms, statistics, visualization, database-oriented or data warehouse-oriented, etc. The

    classification can also take into account the degree of user interaction involved in the datamining process such as query-driven systems, interactive exploratory systems, or

    autonomous systems. A comprehensive system would provide a wide variety of data mining

    techniques to fit different situations and options, and offer different degrees of userinteraction.

    Enterprise Resource planning (ERP)

    ERP is a software architecture that facilitates the flow of information among the different functions

    within an enterprise. Similarly, ERP facilitates information sharing across organizational units andgeographical locations.3 It enables decision-makers to have an enterprise-wide view of the

    information they need in a timely, reliable and consistent fashion.

    ERP provides the backbone for an enterprise-wide information system. At the core of this enterprisesoftware is a central4 database which draws data from and feeds data into modular applications that

    operate on a common computing platform, thus standardizing business processes and data

    definitions into a unified environment. With an ERP system, data needs to be entered only once. Thesystem provides consistency and visibilityor transparencyacross the entire enterprise. A primarybenefit of ERP is easier access to reliable, integrated information. A related benefit is the elimination

    of redundant data and the rationalization of processes, which result in substantial cost savings. The

    integration among business functions facilitates communication and information sharing, leading todramatic gains in productivity and speed.

    The Components of an ERP System - The components of an ERP system are the commoncomponents of a Management Information System (MIS).

    ERP Software - Module based ERP software is the core of an ERP system. Each software

    module automates business activities of a functional area within an organization. Common

    ERP software modules include product planning, parts purchasing, inventory control,product distribution, order tracking, finance, accounting and human resources aspects of an

    organization.

    Business Processes - Business processes within an organization falls into three levels -strategic planning, management control and operational control. ERP has been promoted as

    solutions for supporting or streamlining business processes at all levels. Much of ERP

    success, however, has been limited to the integration of various functional departments.

    ERP Users - The users of ERP systems are employees of the organization at all levels, from

    workers, supervisors, mid-level managers to executives.

  • 7/30/2019 ISM Notes Part II

    20/20

    Hardware and Operating Systems - Many large ERP systems are UNIX based. Windows

    NT and Linux are other popular operating systems to run ERP software. Legacy ERP

    systems may use other operating systems.

    The Boundary of an ERP System - The boundary of an ERP system is usually small than the

    boundary of the organization that implements the ERP system. In contrast, the boundary of supply

    chain systems and ecommerce systems extends to the organization's suppliers, distributors, partnersand customers. In practice, however, many ERP implementations involve the integration of ERP

    with external information systems.

    ERP vs. CRM and SCM

    CRM (Customer Relationship Management) and SCM (Supply Chain Management) are two other

    categories of enterprise software that are widely implemented in corporations and non-profit

    organizations. While the primary goal of ERP is to improve and streamline internal businessprocesses, CRM attempts to enhance the relationship with customers and SCM aims to facilitate the

    collaboration between the organization, its suppliers, the manufacturers, the distributors and the

    partners.


Recommended