Below are the major changes in BI 7.0
or 2004S version when compared with earlier versions.
- In Infosets now you can include Infocubes as well.
- The Remodeling transaction helps you add new key figure and characteristics and handles historical data as well without much hassle. This is only for info cube.
- The BI accelerator (for now only for infocubes) helps in reducing query run time by almost a factor of 10 - 100. This BI accl is a separate box and would cost more. Vendors for these would be HP or IBM.
- The monitoring has been imprvoed with a new portal based cockpit. Which means you would need to have an EP guy in ur project for implementing the portal !
- Search functionality hass improved!! You can search any object. Not like 3.x.
- Transformations are in and routines are passe! Yess, you can always revert to the old transactions too.
- Renamed ODS as DataStore Object(DSO).
- Inclusion of Write-optmized DataStore which does not have any change log and the requests do need any activation
- Unification of Transfer and Update rules
- Introduction of "end routine" and "Expert Routine"
- Push of XML data into BI system (into PSA) without Service API or Delta Queue
- Intoduction of BI accelerator that significantly improves the performance.
- Load through PSA has become a must. I am not too sure about this. It looks like we would not have the option to bypass the PSA during data load.
Metadata Search (Developer Functionality) :
- It is possible to search BI metadata (such as InfoCubes, InfoObjects, queries, Web templates) using the TREX search engine. This search is integrated into the Metadata Repository, the Data Warehousing Workbench and to some degree into the object editors. With the simple search, a search for one or all object types is performed in technical names and in text.
- During the text search, lower and uppercase are ignored and the object will also be found when the case in the text is different from that in the search term. With the advanced search, you can also search in attributes. These attributes are specific to every object type. Beyond that, it can be restricted for all object types according to the person who last changed it and according to the time of the change.
- For example, you can search in all queries that were changed in the last month and that include both the term "overview" in the text and the characteristic customer in the definition. Further functions include searching in the delivered (A) version, fuzzy search and the option of linking search terms with “AND” and “OR”.
- "Because the advanced search described above offers more extensive options for search in metadata, the function ""Generation of Documents for Metadata"" in the administration of document management (transaction RSODADMIN) was deleted. You have to schedule (delta) indexing of metadata as a regular job (transaction RSODADMIN).
- Effects on Customizing
- Installation of TREX search engine
- Creation of an RFC destination for the TREX search engine
- Entering the RFC destination into table RSODADMIN_INT
- Determining relevant object types
- Initial indexing of metadata"
Remote Activation of DataSources (Developer Functionality)
:
When
activating Business Content in BI, you can activate DataSources remotely from
the BI system. This activation is subject to an authorization check. You need
role SAP_RO_BCTRA. Authorization object S_RO_BCTRA is checked. The
authorization is valid for all DataSources of a source system. When the objects
are collected, the system checks the authorizations remotely, and issues a
warning if you lack authorization to activate the DataSources.
In BI, if you trigger the transfer of the Business Content in the active
version, the results of the authorization check are based on the cache. If you
lack the necessary authorization for activation, the system issues a warning
for the DataSources. BW issues an error for the corresponding
source-system-dependent objects (transformations, transfer rules, transfer
structure, InfoPackage, process chain, process variant). In this case, you can
use Customizing for the extractors to manually transfer the required
DataSources in the source system from the Business Content, replicate them in
the BI system, and then transfer the corresponding source-system-dependent
objects from the Business Content. If you have the necessary authorizations for
activation, the DataSources in the source system are transferred to the active
version and replicated in the BI system. The source-system-dependent objects
are activated in the BI system.
Source
systems and/or BI systems have to have BI Service API SAP NetWeaver 2004s at
least; otherwise remote activation is not supported. In this case, you have to
activate the DataSources in the source system manually and then replicate them
to the BI system.
Copy Process Chains (Developer Functionality):
You find this function in the Process Chain menu and use it to copy the process chain you have selected, along with its references to process variants, and save it under a new name and description.
You find this function in the Process Chain menu and use it to copy the process chain you have selected, along with its references to process variants, and save it under a new name and description.
InfoObjects in Hierarchies (Data Modeling):
1 Up to
Release SAP NetWeaver 2004s, it was not possible to use InfoObjects with a
length longer than 32 characters in hierarchies. These types of InfoObjects
could not be used as a hierarchy basic characteristic and it was not possible
to copy characteristic values for such InfoObjects as foreign characteristic
nodes into existing hierarchies. From SAP NetWeaver 2004s, characteristics of
any length can be used for hierarchies.
To
load hierarchies, the PSA transfer method has to be selected (which is always
recommended for loading data anyway). With the IDOC transfer method, it
continues to be the case that only hierarchies can be loaded that contain
characteristic values with a length of less than or equal to 32 characters.
Parallelized Deletion of Requests in DataStore Objects
(Data Management) :
Now you can delete active requests in a DataStore object
in parallel. Up to now, the requests were deleted serially within an LUW. This
can now be processed by package and in parallel.
Object-Specific Setting of the Runtime
Parameters of DataStore Objects (Data Management):
Now you can set the runtime parameters of DataStore objects by
object and then transport them into connected systems. The following parameters
can be maintained:
Ø
Package
size for activation
Ø
Package
size for SID determination
Ø
Maximum
wait time before a process is designated lost
Ø
Type
of processing: Serial, Parallel(batch), Parallel (dialog)
Ø
Number
of processes to be used
Ø
Server/server
group to be used
Enhanced Monitor for Request Processing in DataStore
Objects (Data Management):
1 For
the request operations executed on DataStore objects (activation, rollback and
so on), there is now a separate, detailed monitor. In previous releases,
request-changing operations are displayed in the extraction monitor. When the
same operations are executed multiple times, it will be very difficult to
assign the messages to the respective operations.
In
order to guarantee a more simple error analysis and optimization potential
during configuration of runtime parameters, as of release SAP NetWeaver 2004s,
all messages relevant for DataStore objects are displayed in their own monitor.
Write-Optimized DataStore Object (Data Management):
1 Up to
now it was necessary to activate the data loaded into a DataStore object to
make it visible to reporting or to be able to update it to further
InfoProviders. As of SAP NetWeaver 2004s, a new type of DataStore object is
introduced: the write-optimized DataStore object.
The
objective of the new object type is to save data as efficiently as possible in
order to be able to further process it as quickly as possible without addition
effort for generating SIDs, aggregation and data-record based delta. Data that
is loaded into write-optimized DataStore objects is available immediately for
further processing. The activation step that has been necessary up to now is no
longer required.
The
loaded data is not aggregated. If two data records with the same logical key
are extracted from the source, both records are saved in the DataStore object.
During loading, for reasons of efficiency, no SID values can be determined for
the loaded characteristics. The data is still available for reporting. However,
in comparison to standard DataStore objects, you can expect to lose performance
because the necessary SID values have to be determined during query runtime.
Deleting from the Change Log (Data Management):
The Deletion of Requests from the Change Log process type supports the deletion of change log files. You select DataStore objects to determine the selection of requests. The system supports multiple selections. You select objects in a dialog box for this purpose. The process type supports the deletion of requests from any number of change logs.
The Deletion of Requests from the Change Log process type supports the deletion of change log files. You select DataStore objects to determine the selection of requests. The system supports multiple selections. You select objects in a dialog box for this purpose. The process type supports the deletion of requests from any number of change logs.
Using InfoCubes in InfoSets (Data Modeling):
1 You
can now include InfoCubes in an InfoSet and use them in a join. InfoCubes are
handled logically in InfoSets like DataStore objects. This is also true for
time dependencies. In an InfoCube, data that is valid for different dates can
be read.
For
performance reasons you cannot define an InfoCube as the right operand of a
left outer join. SAP does not generally support more than two InfoCubes in an
InfoSet.
Pseudo Time Dependency of DataStore Objects and InfoCubes
in InfoSets (Data Modeling) :
In BI only master data can be defined as a time-dependent data source. Two additional fields/attributes are added to the characteristic. DataStore objects and InfoCubes that are being used as InfoProviders in the InfoSet cannot be defined as time dependent. As of SAP NetWeaver 2004s, you can specify a date or use a time characteristic with DataStore objects and InfoCubes to describe the validity of a record. These InfoProviders are then interpreted as time-dependent data sources.
In BI only master data can be defined as a time-dependent data source. Two additional fields/attributes are added to the characteristic. DataStore objects and InfoCubes that are being used as InfoProviders in the InfoSet cannot be defined as time dependent. As of SAP NetWeaver 2004s, you can specify a date or use a time characteristic with DataStore objects and InfoCubes to describe the validity of a record. These InfoProviders are then interpreted as time-dependent data sources.
Left Outer: Include Filter Value in On-Condition
(Data Modeling) :
- The global properties in InfoSet maintenance have been enhanced by one setting Left Outer: Include Filter Value in On-Condition. This indicator is used to control how a condition on a field of a left-outer table is converted in the SQL statement.
This affects the query results:
Ø
If
the indicator is set, the condition/restriction is included in the on-condition
in the SQL statement. In this case the condition is evaluated before the join
Ø
If
the indicator is not set, the condition/restriction is included in the
where-condition. In this case the condition is only evaluated after the join
Ø
The
indicator is not set by default
Key Date Derivation from Time Characteristics (Data
Modeling) :
Key dates can be derived from the time characteristics 0CALWEEK, 0CALMONTH, 0CALQUARTER, 0CALYEAR, 0FISCPER, 0FISCYEAR: It was previously possible to specify the first, last or a fixed offset for key date derivation. As of SAP NetWeaver 2004s, you can also use a key date derivation type to define the key date.
Key dates can be derived from the time characteristics 0CALWEEK, 0CALMONTH, 0CALQUARTER, 0CALYEAR, 0FISCPER, 0FISCYEAR: It was previously possible to specify the first, last or a fixed offset for key date derivation. As of SAP NetWeaver 2004s, you can also use a key date derivation type to define the key date.
Repartitioning of InfoCubes and DataStore Objects (Data
Management):
With SAP NetWeaver 2004s, the repartitioning of InfoCubes and DataStore objects on the database that are already filled is supported. With partitioning, the runtime for reading and modifying access to InfoCubes and DataStore objects can be decreased. Using repartitioning, non-partitioned InfoCubes and DataStore objects can be partitioned or the partitioning schema for already partitioned InfoCubes and DataStore objects can be adapted.
With SAP NetWeaver 2004s, the repartitioning of InfoCubes and DataStore objects on the database that are already filled is supported. With partitioning, the runtime for reading and modifying access to InfoCubes and DataStore objects can be decreased. Using repartitioning, non-partitioned InfoCubes and DataStore objects can be partitioned or the partitioning schema for already partitioned InfoCubes and DataStore objects can be adapted.
Remodeling InfoProviders (Data Modeling):
- As of SAP NetWeaver 2004s, you can change the structure of InfoCubes into which you have already loaded data, without losing the data.
You have the following remodeling
options:
- For characteristics:
Ø
Inserting,
or replacing characteristics with: Constants, Attribute of an InfoObject within
the same dimension, Value of another InfoObject within the same dimension,
Customer exit (for user-specific coding)
Ø
Delete
- For key figures:
Ø
Inserting:
Constants, Customer exit (for user-specific coding).
Ø
Replacing
key figures with: Customer exit (for user-specific coding).
Ø Delete
- SAP NetWeaver 2004s does not support the remodeling of InfoObjects or DataStore objects. This is planned for future releases.
- Before you start remodeling, make sure:
A. You
have stopped any process chains that run periodically and affect the
corresponding InfoProvider. Do not restart these process chains until
remodeling is finished.
B. There
is enough available tablespace on the database.
- After remodeling, check which BI objects that are connected to the InfoProvider (transformation rules, MultiProviders, queries and so on) have been deactivated. You have to reactivate these objects manually.
Parallel Processing for Aggregates (Performance):
1
The change run, rollup, condensing and checking up multiple aggregates can be executed in parallel. Parallelization takes place using the aggregates. The parallel processes are continually executed in the background, even when the main process is executed in the dialog.
The change run, rollup, condensing and checking up multiple aggregates can be executed in parallel. Parallelization takes place using the aggregates. The parallel processes are continually executed in the background, even when the main process is executed in the dialog.
This
can considerably decrease execution time for these processes. You can determine
the degree of parallelization and determine the server on which the processes
are to run and with which priority.
If no
setting is made, a maximum of three processes are executed in parallel. This
setting can be adjusted for a single process (change run, rollup, condensing of
aggregates and checks). Together with process chains, the affected setting can
be overridden for every one of the processes listed above. Parallelization of
the change run according to SAP Note 534630 is obsolete and is no longer being
supported.
Multiple Change Runs (Performance):
1 You
can start multiple change runs simultaneously. The prerequisite for this is
that the lists of the master data and hierarchies to be activated are different
and that the changes affect different InfoCubes. After a change run, all
affected aggregates are condensed automatically.
If a
change run terminates, the same change run must be started again. You have to
start the change run with the same parameterization (same list of
characteristics and hierarchies). SAP Note 583202 is obsolete.
Partitioning Optional for Aggregates (Performance):
Up to
now, the aggregate fact tables were partitioned if the associated InfoCube was
partitioned and the partitioning characteristic was in the aggregate. Now it is
possible to suppress partitioning for individual aggregates. If aggregates do
not contain much data, very small partitions can result. This affects read
performance. Aggregates with very little data should not be partitioned.
Aggregates that are not to be partitioned have to be activated and filled again
after the associated property has been set.
MOLAP Store (Deleted) (Performance):
Previously you were able to
create aggregates either on the basis of a ROLAP store or on the basis of a
MOLAP store. The MOLAP store was a platform-specific means of optimizing query
performance. It used Microsoft Analysis Services and, for this reason, it was
only available for a Microsoft SQL server database platform. Because HPA
indexes, available with SAP NetWeaver 2004s, are a platform-independent
alternative to ROLAP aggregates with high performance and low administrative
costs, the MOLAP store is no longer being supported.
Data Transformation (Data Management):
A
transformation has a graphic user interfaces and replaces the transfer rules
and update rules with the functionality of the data transfer process (DTP).
Transformations are generally used to transform an input format into an output
format. A transformation consists of rules. A rule defines how the data content
of a target field is determined. Various types of rule are available to the
user such as direct transfer, currency translation, unit of measure conversion,
routine, read from master data.
Block transformations can be realized using different data package-based rule
types such as start routine, for example. If the output format has key fields,
the defined aggregation behavior is taken into account when the transformation
is performed in the output format. Using a transformation, every (data) source
can be converted into the format of the target by using an individual
transformation (one-step procedure). An InfoSource is only required for complex
transformations (multistep procedures) that cannot be performed in a one-step
procedure.
The
following functional limitations currently apply:
You
cannot use:
Ø
hierarchies
as the source or target of a transformation
Ø
use
master data as the source of a transformation
Ø
use
a template to create a transformation
No
documentation has been created in the metadata repository yet for
transformations.
transformations.
In the
transformation there is no check for referential integrity, the InfoObject
transfer routines are not considered and routines cannot be created using the return table.
transfer routines are not considered and routines cannot be created using the return table.
Quantity Conversion :
As of SAP NetWeaver 2004s you can create quantity conversion
types using transaction RSUOM. The business transaction rules of the conversion
are established in the quantity conversion type. The conversion type is a
combination of different parameters (conversion factors, source and target
units of measure) that determine how the conversion is performed. In terms of
functionality, quantity conversion is structured similarly to currency
translation. Quantity conversion allows you to convert key figures with units
that have different units of measure in the source system into a uniform unit
of measure in the BI system when you update them into InfoCubes.
Data Transfer Process :
You use the data transfer process (DTP) to transfer data
within BI from a persistent object to another object in accordance with certain
transformations and filters. In this respect, it replaces the InfoPackage,
which only loads data to the entry layer of BI (PSA), and the data mart
interface. The data transfer process makes the transfer processes in the data
warehousing layer more transparent. Optimized parallel processing improves the
performance of the transfer process (the data transfer process determines the
processing mode). You can use the data transfer process to separate delta
processes for different targets and you can use filter options between the
persistent objects on various levels. For example, you can use filters between
a DataStore object and an InfoCube. Data transfer processes are used for
standard data transfer, for real-time data acquisition, and for accessing data
directly. The data transfer process is available as a process type in process
chain maintenance and is to be used in process chains.
ETL Error Handling :
The data transfer process supports you in handling data
records with errors. The data transfer process also supports error handling for
DataStore objects. As was previously the case with InfoPackages, you can
determine how the system responds if errors occur. At runtime, the incorrect
data records are sorted and can be written to an error stack (request-based
database table). After the error has been resolved, you can further update data
to the target from the error stack. It is easier to restart failed load
processes if the data is written to a temporary store after each processing
step. This allows you to determine the processing step in which the error
occurred. You can display the data records in the error stack from the monitor
for the data transfer process request or in the temporary storage for the
processing step (if filled). In data transfer process maintenance, you
determine the processing steps that you want to store temporarily.
InfoPackages :
InfoPackages only load the data into the input layer of
BI, the Persistent Staging Area (PSA). Further distribution of the data within
BI is done by the data transfer processes. The following changes have occurred
due to this:
Ø
New
tab page: Extraction -- The Extraction tab page includes the settings for
adaptor and data format that were made for the DataSource. If data transfer
from files occurred, the External Data tab page is obsolete; the settings are
made in DataSource maintenance
Ø Tab
page: Processing -- Information on how the data is updated is obsolete because
further processing of the data is always controlled by data transfer processes
Ø Tab
page: Updating -- On the Updating tab page, you can set the update mode to the
PSA depending on the settings in the DataSource. In the data transfer process,
you now determine how the update from the PSA to other targets is performed.
Here you have the option to separate delta transfer for various targets
For real-time acquisition with the Service API, you create
special InfoPackages in which you determine how the requests are handled by the
daemon (for example, after which time interval a request for real-time data
acquisition should be closed and a new one opened). For real-time data
acquisition with Web services (push), you also create special InfoPackages to
set certain parameters for real-time data acquisition such as sizes and time
limits for requests.
PSA :
The persistent staging area (PSA), the entry layer for
data in BI, has been changed in SAP NetWeaver 2004s. Previously, the PSA table
was part of the transfer structure. You managed the PSA table in the Administrator
Workbench in its own object tree. Now you manage the PSA table for the entry
layer from the DataSource. The PSA table for the entry layer is generated when
you activate the DataSource. In an object tree in the Data Warehousing
Workbench, you choose the context menu option Manage to display a DataSource in
PSA table management. You can display or delete data here. Alternatively, you
can access PSA maintenance from the load process monitor. Therefore, the PSA
tree is obsolete.
Real-Time Data Acquisition :
Real-time data acquisition supports tactical decision
making. You use real-time data acquisition if you want to transfer data to BI
at frequent intervals (every hour or minute) and access this data in reporting
frequently or regularly (several times a day, at least). In terms of data
acquisition, it supports operational reporting by allowing you to send data to
the delta queue or PSA table in real time. You use a daemon to transfer
DataStore objects that have been released for reporting to the ODS layer at
frequent regular intervals. The data is stored persistently in BI. You can use
real-time data acquisition for DataSources in SAP source systems that have been
released for real time, and for data that is transferred into BI using the Web
service (push). A daemon controls the transfer of data into the PSA table and
its further posting into the DataStore object. In BI, InfoPackages are created
for real-time data acquisition. These are scheduled using an assigned daemon
and are executed at regular intervals. With certain data transfer processes for
real-time data acquisition, the daemon takes on the further posting of data to
DataStore objects from the PSA. As soon as data is successfully posted to the
DataStore object, it is available for reporting. Refresh the query display in
order to display the up-to-date data. In the query, a time stamp shows the age
of the data. The monitor for real-time data acquisition displays the available
daemons and their status. Under the relevant DataSource, the system displays
the InfoPackages and data transfer processes with requests that are assigned to
each daemon. You can use the monitor to execute various functions for the
daemon, DataSource, InfoPackage, data transfer process, and requests.
Archiving Request Administration Data :
You can now archive log and administration data requests.
This allows you to improve the performance of the load monitor and the monitor
for load processes. It also allows you to free up tablespace on the database.
The archiving concept for request administration data is based on the SAP
NetWeaver data archiving concept. The archiving object BWREQARCH contains
information about which database tables are used for archiving, and which
programs you can run (write program, delete program, reload program). You
execute these programs in transaction SARA (archive administration for an
archiving object). In addition, in the Administration functional area of the
Data Warehousing Workbench, in the archive management for requests, you can
manage archive runs for requests. You can execute various functions for the
archive runs here.
After an upgrade, use BI background management or transaction
SE38 to execute report RSSTATMAN_CHECK_CONVERT_DTA and report
RSSTATMAN_CHECK_CONVERT_PSA for all objects (InfoProviders and PSA tables).
Execute these reports at least once so that the available request information
for the existing objects is written to the new table for quick access, and is
prepared for archiving. Check that the reports have successfully converted your
BI objects. Only perform archiving runs for request administration data after
you have executed the reports.
Flexible process path based on multi-value decisions :
The workflow and decision process types support the event
Process ends with complex status. When you use this process type, you can
control the process chain process on the basis of multi-value decisions. The
process does not have to end simply successfully or with errors; for example,
the week day can be used to decide that the process was successful and
determine how the process chain is processed further. With the workflow option,
the user can make this decision. With the decision process type, the final
status of the process, and therefore the decision, is determined on the basis
of conditions. These conditions are stored as formulas.
Evaluating the output of system commands :
You use this function to decide
whether the system command process is successful or has errors. You can do this
if the output of the command includes a character string that you defined. This
allows you to check, for example, whether a particular file exists in a
directory before you load data to it. If the file is not in the directory, the
load process can be repeated at pre-determined intervals.
Repairing and repeating process chains :
You use this function to repair processes that were
terminated. You execute the same instance again, or repeat it (execute a new
instance of the process), if this is supported by the process type. You call
this function in log view in the context menu of the process that has errors.
You can restart a terminated process in the log view of process chain
maintenance when this is possible for the process type.
If the process cannot be repaired or repeated after termination,
the corresponding entry is missing from the context menu in the log view of
process chain maintenance. In this case, you are able to start the subsequent
processes. A corresponding entry can be found in the context menu for these
subsequent processes.
Executing process chains synchronously :
You use this function to schedule and execute the process
in the dialog, instead of in the background. The processes in the chain are
processed serially using a dialog process. With synchronous execution, you can
debug process chains or simulate a process chain run.
Error handling in process chains :
You use this function in the attribute maintenance of a
process chain to classify all the incorrect processes of the chain as
successful, with regard to the overall status of the run, if you have scheduled
a successor process Upon Errors or Always. This function is relevant if you are
using metachains. It allows you to continue processing metachains despite
errors in the subchains, if the successor of the subchain is scheduled Upon Success.
Determining the user that executes the process chain :
You use this function in the attribute maintenance of a
process chain to determine which user executes the process chain. In the
default setting, this is the BI background user.
Display mode in process chain maintenance :
When you access process chain maintenance, the process
chain display appears. The process chain is not locked and does not call the
transport connection. In the process chain display, you can schedule without
locking the process chain.
Checking the number of background processes available for
a process chain :
During the check, the system calculates the number of
parallel processes according to the structure of the tree. It compares the
result with the number of background processes on the selected server (or the
total number of all available servers if no server is specified in the
attributes of the process chain). If the number of parallel processes is
greater than the number of available background processes, the system highlights
every level of the process chain where the number of processes is too high, and
produces a warning.
Open Hub / Data Transfer Process Integration :
As of SAP NetWeaver 2004s SPS 6, the open hub destination
has its own maintenance interface and can be connected to the data transfer
process as an independent object. As a result, all data transfer process
services for the open hub destination can be used. You can now select an open
hub destination as a target in a data transfer process. In this way, the data
is transformed as with all other BI objects. In addition to the InfoCube,
InfoObject and DataStore object, you can also use the DataSource and InfoSource
as a template for the field definitions of the open hub destination. The open
hub destination now has its own tree in the Data Warehousing Workbench under
Modeling. This tree is structured by InfoAreas.
The open hub service with the InfoSpoke that was provided until
now can still be used. We recommend, however, that new objects are defined with
the new technology.
No comments:
Post a Comment