Automating Data Warehouse Conceptual Schema Design And Evaluation

1y ago
13 Views
2 Downloads
732.82 KB
10 Pages
Last View : 22d ago
Last Download : 3m ago
Upload by : Camille Dion
Transcription

Automating Data Warehouse Conceptual Schema Design and EvaluationCassandra PhippsKaren C. DavisABB Inc.650 Ackerman Rd.Columbus, OH 43202casie.phipps@us.abb.comECECS Dept.University of CincinnatiCincinnati, OH 45221-0030karen.davis@uc.eduAbstractgathering defines the requirements of a data warehouseusing source data or a schema from an OLTP system.The benefits of using a source-driven requirementsgathering approach are that minimal user time is requiredto start the project, and complete data is supplied since theexisting data in the OLTP system provides the frameworkfor the data warehouse. A disadvantage of this approachis that incomplete knowledge of user needs and relianceon only the OLTP database may not produce an adequatedata warehouse schema. To alleviate this disadvantage weinclude an opportunity for user refinement in our designprocess. User-driven requirements gathering is a methodbased on investigating functions users perform. We focuson user needs as represented in a set of queries to beapplied against the data warehouse. Having user-drivenrequirements prevents critical business needs from beingoverlooked. To illustrate the proposed algorithms, bothsource-driven and user-driven requirements are gatheredfrom the TPC-H Benchmark [TPC99]. We treat thebenchmark schema as an OLTP schema for sourcederivation and its queries as user requirements.Section 2 discusses data models used to representschemas and gives a brief overview of related work indata warehouse design. Section 3 presents and illustratesan automated approach to constructing candidateconceptual schemas for a data warehouse from an OLTPschema. Section 4 gives an algorithm for evaluatingcandidate schemas based on user requirements as well asa guideline for manual schema refinement. Section 5discusses advantages and known limitations of our workalong with topics for future work.The popularity of data warehouses for analysis of data hasgrown tremendously, but much of the creation of datawarehouses is done manually. We propose and illustratealgorithms for automatic conceptual schema development andevaluation. Our creation algorithm uses an enterprise schemaof an operational database as a starting point for source-drivendata warehouse schema design. Candidate conceptual schemasare created using the ME/R model, extended to note whereadditional user input can be used to further refine a schema.Our evaluation algorithm follows a user-driven requirementsapproach that utilizes queries to guide selection of candidateschemas most likely to meet user needs. In addition, we proposea guideline of manual steps to refine a conceptual schema to suitadditional user needs, for example, the level of detail needed fordate fields. The algorithms are illustrated using the TPC-HBenchmark schema and queries. Our algorithms provide afoundation for a software tool to create and evaluate datawarehouse conceptual schemas.1. IntroductionAs conventional transaction processing systems havematured, becoming faster and more stable, the focus oforganizational needs has changed. Increasing the value oftransaction processing systems “means turning data intoactionable information” [R96a]. Although traditionalOnLine Transaction Processing (OLTP) systems mayhave some, or all, of the necessary data, it is not easilyaccessed by the user for analytical processing. The needfor OnLine Analytical Processing (OLAP) gives rise tothe data warehouse concept.A data warehouse creation process consists of fivesteps: pre-development activities, architecture selection,schema creation, warehouse population, and datawarehouse maintenance [M97, SC99]. The focus of thispaper is the schema creation phase and its automation;while this phase includes conceptual, logical, and physicalschema design, we only address conceptual design of adata warehouse here. The conceptual model allows ahigh-level design of entities and their relationships,represented in a user-friendly manner independent ofimplementation issues. A conceptual schema is adescription of the data to be in the data warehouse that isunderstandable by end users to verify requirements,identify possible gaps, and conduct analysis for businessgoals.We propose automated techniques to develop andevaluate candidate data warehouse conceptual schemasusing source-driven and user-driven requirementsgathering, respectively.Source-driven requirements2. Related WorkThe goal of creating a conceptual schema isto translate user requirements into an abstractrepresentation understandable to the user, that isindependent of implementation issues, but isformal and complete, so that it can betransformed into the next logical schemawithout ambiguities [TB99].Requirements of a conceptual model for a data warehouseinclude a) providing a modeling construct to representbusiness facts and their properties, b) connecting temporaldimensions to facts, c) relating objects with theirproperties and associations, d) defining relationshipsbetween objects and business facts, and e) outliningdimensions and their respective hierarchies [TB99]. We1

discuss our choice for a conceptual model and relatedautomated approaches below.2.1 Conceptual Model SelectionIn the OLTP arena, Entity/Relationship-based modelsare most widely used, and in the OLAP arena dimensionalmodeling is most popular. An ER schema is a graphicalrepresentation of entities and their relationships to eachother. ER-based models have been extended to includethe dimensional functionality necessary in datawarehousing [B99, FS99, TB99, HS00]. Dimensionalmodels organize data based on business rules of anorganization. Dimensional models are based on the ideathat an organization’s facts are central and the dataunfolds around it. The most well-known dimensionalmodel is the Star model [BH98]; other variations are theSnowflake model [BH98] and the Dimensional FactModel [GM98a]. Further discussion about ER versusdimensional modeling [K95, R96b] and converting froman ER model to a dimensional model [K97, M98] is givenelsewhere.Conceptual and logical schemas for data warehousescan be built with ER-based or dimensional-based models.It is possible that the conceptual schema may be in oneform and the logical another. For example, McGuff[M98] uses ER modeling for conceptual schemas anddimensional for the logical and physical designs, whileWu and Buchmann [WB97] do the opposite for schemacreation. Table 1 gives a summary of models used byvarious authors for conceptual and logical schemacreation. The Star model is popular for both conceptualand logical modeling, but may not be the best choice forend users since it does not show drill-down and roll-uppaths [BH98].We choose the ME/R model forconceptual schema creation and the Star model for logicalschema creation. Translating from ME/R to Star form isstraightforward; our algorithm and examples are notprovided in this paper due to space limitations.The ME/R (Multidimensional Entity-Relationship)model is similar to, but simpler than, the EVER andStarER models. Figure 1 is an ME/R example of acustomer orders application. The central construct is afact node, represented by a diamond, that containsbusiness measures. A level node, represented by arectangle, contains information about the measures. Eachlevel or fact node has attributes, represented as ellipses,that are descriptors for this entity. These are the threemain constructs that make up the ME/R model, and theyare connected by various relationships. In the example,Order is the fact node. Product, Customer, Day, Month,Week, and Year are level nodes. A fact node is related toa level by a dimension edge that is an undirected line.The has edge connects facts or levels to their attributesand is also an undirected line. The classification edge isthe relationship between levels. This is represented by apitchfork-like symbol at one end of a directed line. Allthe levels of the fact node, related by classification edges,represent a dimension of the model. The dimension thatrepresents the order date in our example is made up of aDay level, a classification of the Month level that in turnis a classification level of Year. The directed edgebetween the levels is important in that it showshierarchical aggregation. This is especially useful when alevel node can be a classification node for more than oneother level as seen in the Day to Week relationship. In ourexample, we can roll-up from the Day level to the Monthlevel, changing the level of aggregation shown for the factattributes. Week is an alternate roll-up path from tual Design[M98][B99][TB99][HS00][FS99][TK01][WB97, R96b, K96b,BH98, M94, M98]Logical Design[BH98][K96b, BE99, CD97,K97, BH98, M98][BE99, CD97, BH98][GR98, GR99]Table 1. Models for Conceptual and Logical SchemasFigure 1. An Example ME/R SchemaWith only three main constructs, the ME/R model issimple for users to read, and aggregation is shownthrough the hierarchy of levels representing a dimension.This model resembles the Star model, leading to ease ofunderstanding by both users and logical modelers of thedata warehouse. Details of additional ME/R modelingconstructs are given elsewhere [HS00].2

2.2 Automated Approaches to Schema DesignThe initial determination of measures, facts, or eventscan prove to be the most difficult part of the designprocess, and is usually done manually.Differentapproaches to finding measures are suggested: (1)deriving the fact table of a Star schema by selecting the“many-to-many relationships in the ER model containingnumeric and additive nonkey facts” [K97], (2) findingcandidate measures by analyzing the business queries fordata items indicating the performance of the business[BH98], (3) finding facts based on entities mostfrequently updated [GM98b], and (4) observing that “factproperties are usually numerical data, and can besummarized (or aggregated)” [TB99]. The third approachis implemented in a CASE tool [GR01]; we include thelast approach in our algorithm in Section 3.There have been some proposals for semi-automatedor automated approaches to schema creation from OLTPsystems.Automating logical schema creation andphysical schema design have been proposed [TS97,CT98, GR98, BE99, TS99, S99, HL00, HS00, MK00].None of the previous approaches include an automatedmechanism for finding candidate measures. Our researchis the first effort that addresses automation of creating anentire candidate conceptual schema from an OLTPschema, including the initial determination of facts ormeasures.3. Conceptual Schema CreationOur approach to schema generation has two basicpremises. One is that numeric fields represent measures ofpotential interest to a business and the more numericfields in an entity the more likely it is that the entity is anevent or fact. The second premise is that the cardinality ofa relationship determines how useful a related entity is tothe schema being created. Any entity related with a manyrelationship is of likely importance (and any of its entitiesmay be as well), since it is potentially a higher level ofabstraction.Our algorithm for creating candidate conceptualschemas has five steps that result in candidate schemascentered around likely business events, with relationshipsto other entities forming the dimensions describing theseevents. Although there are instances where the eventschosen have little or no meaning to users and additionalcommonly used calculations may need to be added, thebulk of the initial schema creation is automated. Schemascan be manually refined as described in Section 4.Input to the algorithm is an ER schema representedhere in table data structures. The steps of the algorithmare generalized as follows:1. Find entities with numeric fields and create a factnode for each entity identified.2. Create numeric attributes of each fact node, basedon numeric fields in the entities.3. Create date and or time levels (dimensions) with anydate/time type fields per fact node.4. Create a level (dimension) containing the remainingentity attributes (non-numeric, non-key, and nondate fields).5. Recursively examine the relationships of the entitiesto add additional levels in a hierarchical manner(creating a dimension).The output of the algorithm is a set of candidateME/R schemas in tabular form. Four additional tables arecreated for use in logical schema creation that store factnode names for the various candidate schemas created,the names of each level of the fact nodes and the levelsthat are sub-levels of the level nodes, and the attributes ofthe facts and levels. These tables are not illustrated in thispaper, only the graphical representations of candidateschemas are presented here.The algorithm is given in Figure 2 with thecorresponding steps labeled. The function calls areinformally described here via an example using the TPCH Benchmark schema [TPC99]. The TPC-H Benchmarkis created specifically to benchmark query performance ina data warehouse environment, but here we treat the TPCH schema as an OLTP schema since it resembles one in anormalized format. There are benefits from the decisionto use the TPC-H benchmark. First, it is an industrybenchmark example and thus not biased toward theschema creation algorithm. Second, order entry and/orpart distribution is a common function of a wide range ofbusinesses. Third, and most important for our purposes, aschema for source-driven requirements and queries foruser-driven analysis are given. We reverse engineer anER schema from the TPC-H schema and use it as input toour algorithm.Step 1 orders the entities with numeric fields indescending order of number of numeric fields.Descending order is not necessary but the entities with thegreatest number of numeric fields create candidateschemas that are generally better for answering userqueries. By processing the entities in this order thecandidates that are more likely to be useful are createdfirst. In our example, the only numeric fields are of typedecimal and integer. The result of this step is a list oftables in a ranked order of number of numeric fields, i.e.,LineItem (5), PartSupp (2), Part (2), Orders (2), Supplier(1), and Customer (1), for a total of 6 candidate schemas.Starting with the first entity, LineItem, we create anME/R schema. The fact node is represented by adiamond shape, labeled LineItem Event, shown in Figure3. This fact node becomes the focus of the candidateME/R schema created on this iteration of the loop.In Step 2, the numeric fields for LineItem Event,L LineNumber,L Quantity,L ExtendedPrice,L Discount, and L Tax are added to the diamond asellipses, indicating attributes, as shown in Figure 3.3

Algorithm for Conceptual Schema CreationInput Parameters:Table ColumnsTable RelationsIn/Out Parameters:Fact Node TableFact Attribute TableLevel TableLevel Attribute TableVariables: num tables[]fact nodenum field[]date field[]other field[]// Table containing table name, column name, and column type//for every OLTP table.// Table containing the OLTP schema relationships// Table defining the fact nodes of ME/R schemas and the//OLTP table name that is used to create the fact node.// Table defining the attributes of the fact nodes for the ME/R//schema.// Table defining the levels of the ME/R schema.// Defines the attributes of the levels.// Array of table names from OLTP schema(s) with numeric//fields. Array is ordered in descending order of numeric//fields.// Fact node name.// Array of numeric OLTP attribute field names.// Array of date OLTP attribute field names.// Array of OLTP non-key, non-numeric, non-date/time fields.Method:num tables[] : select order tables numeric fields (Table Columns)for each num tables[j]fact node : create fact node(num tables[j], Fact Node Table)num field[] : select num field (Table Columns, num tables[j])for each num field[m]create fact node attribute (fact node, num field[m], Fact Attribute Table)end for loopdate field[v] : select date field (Table Columns, num tables[j])if isempty(date field[]) thencreate review levels(fact node, Level Table)elsefor each date field[v]create date time level (fact node, date field[v], Level Table)end for loopend ifif exists other field in OLTP table (Table Columns, num tables[j])create level (fact node, num tables[j], Level Table)other field[] : select other fields (Table Columns, num tables[j])for each other field[a])add fields to level (fact node, other field [a], num tables[j], Level Attribute Table)end for loopend ifWalk Relationships (num tables[j], fact node, Table Columns, Table Relations,Level Table, Level Attribute Table)end for loopend algorithm(1)(2)(3)(4)(5)Figure 2. Algorithm for Conceptual Schema CreationFigure 3. LineItem Event with Numeric AttributesNext, as part of Step 3, we identify any date fields ofthis entity. LineItem has three date and/or time fields:L ShipDate, L CommitDate, and L ReceiptDate. Thesefields become the date/time levels or dimensions of ourME/R schema. At this point in the automated process, thegranularity of the date and time dimensions are unknown;user refinement is needed to determine how to representthe levels that make up the date dimension. In ME/Rdiagrams, dimensions are represented by a collection of4

levels in a hierarchical manner. Instead of representingthe date and time dimensions as rectangles normally usedto represent levels in an ME/R diagram, we introduce anew notation, a hexagon, to indicate a portion of theschema where user refinement of date/time granularity isneeded. The ME/R schema created so far as given inFigure 4.Step 4 is only processed if there are attributesremaining in the entity that are not yet processed (i.e., notkey, numeric, or date fields). The remaining fields aregenerally text fields. If there are text fields as part of thisentity then a level (dimension) node is created with thisentity’s name. The level nodes are symbolized byrectangles. Each remaining field becomes an attribute ofthe level node. For our example, LineItem has five suchattributes: L ReturnFlag, L LineStatus, L ShipInstruct,L ShipMode, and L Comment. The new level, LineItem,and its attributes are in Figure 5.Figure 4. Date Levels of LineItem EventStep 5 is the most complicated step, calledWalk Relations, where recursive traversal of relationshipsis conducted. The rest of the TPC-H schema entities(Customer, Orders, Nation, Region, Supplier, Part, andPartSupplier) may be used here. In this step, therelationships of the fact node/event entity are evaluated.Every connected entity is made into a level node with itsattributes. If this level entity is part of the many side ofthe relationship, its relationships are also evaluated andprocessed by calling the sub-procedure again. This stepidentifies the relationship names of interest;Order LineItemandPartSupp LineItem;theserelationships are processed one at a time within the subprocedure loop.The relationship names (Order LineItem andPartSupp LineItem) are used to find the entities that makeup dimensions for LineItem Event. Starting with theOrder LineItem relationship name the first relationship isfound, with the entity Orders. A level node is created forOrders along with its attributes.At this point we have added our first level notderived from the LineItem entity. We now need todetermine if any of the OLTP entities related to Ordersshould be included in this schema. We use relationshipcardinality to indicate how many entity instances arerelated to another entity instance by the relationship. Forthe Orders LineItem relationship for the LineItem entity,the cardinality is on the many side of the relationship withOrders. In other words, one Order can have manyLineItem instances. Walk Relations is recursively calledwith Orders now being the parameter passed. For thesame reason, Customer, Nation, and Region becomelevels in the hierarchy, shown in Figure 5. In theexample, all of the relationships are many-to-one, thus wecontinue to recursively call the Walk Relationsprocedure. If we encounter a relationship that is notmany-to-one or many-to-many, we do not include it in theschema. If the cardinality of Orders to Customer had notbeen a many-to-one relationship, the Customer, Nation,and Region entitities would not have been visited andadded to the schema. As is, the sub-leveling ends withRegion because it has no relationships not alreadyevaluated.The last relationship with LineItem is to PartSupp.This is a many-to-one relationship, so the relationships ofPartSupp are recursively processed as well. PartSupp is alittle different from other entities in this example becauseit has multiple sub-levels. The other difference is that theNation and Region levels already exist from therelationship to the Customer level so that we do not haveto duplicate part of the schema. The first completecandidate schema (for one iteration of the outermost loopin 2) is given in Figure 5.Now that we have completed the first iteration of thealgorithm, we create another candidate schema in thesecond iteration. Steps 1 through 5 are reapplied withPartSupp as the event entity. Since the PartSupp entity ison the one-side of the relationship with the LineItementity, we do not delve any lower on that dimension path.The completed candidate schema is shown in Figure 7.This candidate schema has fewer levels in its dimensions.Only the relationship with the Supplier table leads to sublevels. Another new symbol, the cloud shape, isintroduced to denote that a level may be needed and is notautomatically derived from the OLTP schema. In datawarehouses, events tend to be measured by date or timeperiods. Although the PartSupp entity does not have anydate/time fields, the final design in the data warehouseprobably will. This dimension may be created by thegrain requirements of the user for analysis or by therefresh requirements used to capture data from the OLTPsystem to store into the data warehouse. Dimensionsrepresented by the cloud are added to any fact node withno date or time fields.Following the sequence of the algorithm, the otherentities with numeric fields are examined and candidateschemas are produced. These are given in Figures 7through 10. Not all of the generated conceptual schemasmay prove to be useful to the user; the 6 candidate5

conceptual schemas automatically created by ouralgorithm are evaluated against user requirements in thenext section.Figure 7. Candidate Schema 3: Part EventFigure 5. Candidate Schema 1: LineItem EventFigure 8. Candidate Schema 4: Orders EventFigure 6. Candidate Schema 2: PartSupp Event6

measures that need to be attributes of the fact node toanswer the query. In order to compare the candidateschemas for satisfying the queries, we create a table. Forour example, Table 2 represents the 22 queries and the 6candidate conceptual schemas. In the table, “X” showsthat the candidate schema completely meets the queryrequirement, and “P” means that the schema partiallyanswers the query because a numeric value of interest isnot in a fact node but is in one of the dimensions. A blankentry means that the query is not answerable by thisschema.As an example of how the evaluation algorithmworks, consider query Q1:Figure 9. Candidate Schema 5: Customer EventFigure 10. Candidate Schema 6: Supplier Event4. Candidate Schema Selection and RefinementThe TPC-H Benchmark queries are used here toevaluate which candidate schemas best meet users’ needs.There are two aspects of a query that are used todetermine if a candidate schema can answer a query: thetables in the FROM clause and the numeric fields in theSELECT clause. If a candidate schema does not containthe table(s) in the FROM clause it cannot answer thequery because the fields of that table are not in theschema either. It is unnecessary to check for every field inthe query SELECT statement because the candidateschema generation algorithm dictates that every field in atable of the OLTP system is in the schema. The numericfields from the SELECT clause are essentially theSELECT l returnflag, l linestatus, Sum(l quantity) assum qty, Sum(l extendedprice) as sum base price,Sum(l extendedprice * (1 – l discount)) assum disc price, Sum(l extendedprice * (1 – l discount)* (1 l tax)) as sum charge, Avg(l quantity) asavg qty, Avg(l extendedprice) as avg price,Avg(l discount) as avg disc, Count(*) as count orderFROMlineitemWHERE l shipdate date ‘1998-12-01’ – interval‘[DELTA]’ day(3)GROUP BYl returnflag, l linestatusORDER BYl returnflag, l linestatus;Starting with Candidate Schema 1, since LineItem Eventis the fact node and all numeric fields in the SELECTstatement are attributes of the fact node, this schemasatisfies the query (an “X” entry). Doing the same stepsfor Candidate Schema 2, PartSupp Event, LineItem isagain included in the schema, however, the numeric fieldsare not in the fact node but are in a level node (a “P” entryfor partial match). Candidate Schemas 3 through 6 areprocessed in this manner with only Candidate Schema 4,Orders Event, found to be a partial answer to the query.Evaluating the remaining queries for each schema in thesame manner yields Table 2.The evaluation algorithm identifies whether variouscandidate schemas meet the requirements of the userqueries. The resulting table can be automaticallyprocessed using a graph covering algorithm or manuallyevaluated to determine which candidate schemas to keepand which can be discarded. We describe a manualprocess to illustrate tradeoffs between candidate schemasand motivate our refinement guidelines.There are two types of queries that yield results thatmay require manual refinement before evaluation of thetable: a query with numeric SELECT fields from multipleOLTP entities, and queries with no numeric fields. Anyquery that has numeric SELECT fields from more thanone OLTP table results in only a partial solution to thequery. Q10 is an example of this. Query Q10 has threenumeric fields, L ExtendedPrice, L Discount, andC AcctBal. Because these fields span different OLTPentities, none of our candidate schemas can directlyanswer the query. Any candidate schema that meets theFROM criteria evaluates to a “P” for partial. Manual7

evaluation is needed to determine if the schema canindeed answer the query adequately or whether minorchanges to the candidate schema may allow it tocompletely answer the query. One such change may bethe movement of a numeric field from a level to a factnode.For queries with no numeric fields in the SELECTstatement, the schemas that include all the entities in theFROM statement of the query evaluate as direct answers(i.e., “X”) to the query requirements. This evaluation ofthe candidate schemas is correct because the queries onlyrequire textual or date type information. With norequirement for numerical data all data can come fromany candidate schema with the appropriate levels. QueryQ12 is an example of this type of query. The FROMsection of Query Q12 has two tables, Orders andLineitem. Candidate Schemas 1 and 4 have data fromthese tables and both meet the user requirements of XPXXXPPXPPPPPXTable 2. Candidate Schema EvaluationTable 2 gives us several possibilities for schemas thatmeet our data warehouse requirements as defined by theuser queries. Candidate Schemas 3 and 6 can beeliminated since they do not answer any queries thatcannot be answered by other candidate schemas.Candidate Schema 4, while satisfying many queries, doesnot satisfy any that are not satisfied by Candidate Schema1, and can thus be discarded. Candidate Schema 2 ispromising; it is a stronger schema than Candidate Schema1 for a few of the queries (“X” instead of “P” for Q11 andQ16). Candidate Schema 5 may not be needed, but itanswers Q22 which can only be partially answered byCandidate Schema 1. Further analysis is needed to decideif Candidate Schema 5 can be dropped or is needed aspart of the data warehouse for specific purposes. Theanalysis may result in modifications to the query ormodifications to Candidate Schema 1. Additional steps tofurther refine schemas based on user input or knowledgeof a designer are discussed below.Most of the refinement steps require little knowledgeof the existing OLTP database schema; knowledge of theuser needs is more important. We identify seven manualsteps for conceptual schema refinement (numbering startswith 6 because the automated conceptual design stepsended with 5).6. If user queries are known, eliminate unnecessarycandidate schemas.7. Inspect measures in each fact node. Are theyindeed measures or attributes?8. What is the necessary grain of date/timeinformation?9. Are other calculated fields necessary?10. Can schemas be merged?11. Can any fields be eliminated as not necessary?12. Is there any data required that did not exist in theOLTP database?There may be other additions dictated by need andsituation but the steps above address many of the changesthat need to be made manually. We provide someexamples, but these examples rely on our interpretation ofthe TPC-H information. The following is for illustrativepurposes only and is not meant as a comprehensivesolution for finishing the automatically generatedschemas.Because user queries to be answered by the datawarehouse are known, we are able to eliminate several ofthe candidate schemas in Step 6. The selection of thecandidate schemas is considered a manual step, althoughit is facilitated by the evaluation table, because thedetermination of which schemas to keep is subjective andshould be considered by a designer.Step 7 analyzes the measures of a fact node. Somenumeric fields are actually attributes (descriptors) ratherthan measures. Fields such as width, length, and otherphysical properties are generally attributes used aslimiting criteria in a query rather than as measures of thebusiness. In LineItem Event, the L LineNumber field islikely used to sequence the items that make up the order.L LineNumber is non-additive, and would n

schema design, we only address conceptual design of a data warehouse here. The conceptual model allows a high-level design of entities and their relationships, represented in a user-friendly manner independent of implementation issues. A conceptual schema is a description of the data to be in the data warehouse that is understandable by end .

Related Documents:

HS-PORTAL 150 Hebe-Schiebe-Türbeschlag für Holzelemente Schema-Übersicht und allgemeine Hinweise 3.1.2 Ausführbar mit Führungsschiene HH0130-01/-02 KH0130-01 Schema A Schema D Schema G Schema G-2 Schema G-3 Schema H Schema C Schema F Schema K Schema E Schema L Achtung: Die für den SIEGENIA-AUBI-Beschlag HS-PORTAL 150 angegebenen .

Management under Master Data Define Warehouse Numbers. 2. Check the warehouse number assignment in Customizing for Extended Warehouse Management under Master Data Assign Warehouse Numbers. 3. Check the warehouse number control in Customizing for Extended Warehouse Management under Master Data Define Warehouse Number Control.

support schema evolution [43] to handle data whose structure . by storing FSD as one object without relying on any static schema & E/R model to decompose FSD into relational tables. That is, no schema on write . Flexible schema that is . schema ba

The totality of these behaviors is the graph schema. Drawing a graph schema . The best way to represent a graph schema is, of course, a graph. This is how the graph schema looks for the classic Tinkerpop graph. Figure 2: Example graph schema shown as a property graph . The graph schema is pretty much a property-graph.

location: fort worth, tx warehouse status: approved county: tarrant warehouse capacity: 85,000 warehouse code: 853007 001 location(s) warehouse name: eugene b smith & company , inc license type: unlicensed location: galveston, tx warehouse status: approved county: galveston warehouse capacity: 37,180 warehouse code: 858054 001 location(s)

location: fort worth, tx warehouse status: approved county: tarrant warehouse capacity: 85,000 warehouse code: 853007 001 location(s) warehouse name: eugene b smith & company , inc license type: unlicensed location: galveston, tx warehouse status: approved county: galveston warehouse capacity: 37,180 warehouse code: 858054 001 location(s)

What is a Data Warehouse Architecture? The architecture for a data warehouse consists of more than software and the phYSical storage of data. It should be viewed as an integration of technology and processes. A data warehouse architecture consists of several layers: Conceptual layer - The conceptual layer

Business Accounting Volume 1is the world’s best-selling textbook on bookkeeping and accounting. Now in its tenth edition, it has become the standard introductory text for accounting students and professionals alike. New to this edition: Over 120 brand new review questions for exam practice Coverage of International Accounting Standards 2005 Additional and updated worked examples for areas of .