Science Tools Corporation
Copyright © 1997 - 2024 Science Tools Corporation All rights reserved
Disclaimer
About UsOur ValueProductsConsultingReference Reference Support
 

Questions and Answers

Product Specific Question and Answers
arranged by Topic

1. Staff during the implementation phase

2. Time line feasibility

3. Cost estimate

Additional Details:

4. Extensibility

5. Customized Data-Views


 

1.

Staff during the implementation phase

Question:

How will your staff interface with project management and engineering staff during the implementation phase? Will on-site support staff be required during implementation?

Response:

This will to some extent depend on the scope of our implementation responsibilities. In any case, we do not anticipate that onsite will be required but certainly it will be very helpful and the greater our responsibilities, the more important on-site work will be. We can do substantial work through the internet using secure, encrypted channels and our staff has great telephone skills.

2.

Time line feasibility

Question:

The customer current time line calls for vendor selection in six to nine months, with implementation over the following 12 to 18 months. Is the schedule feasible? Explain.

Response:

The fact that the customer asks this question confirms that time is exceptionally short, yet we know it is possible. The wild-card, from our view, is your staff. One day at the end of April, 1997, when BigSur was just a University project and Science Tools wasn't even a dream, the phone rang. That call brought a visit to the Langley DAAC, which we soon learned was under-the-gun for developing a processing system for data from a satellite set to launch that fall, with a launch-delay deadline of September 1. The Republicans were trying to kill the Mission To Planet Earth and there was real concern any failure would mean the end of it. They had nothing. Their staff was filled with few stars. They wanted a commercial BigSur, on a different operating system, on a different database, written in a different language, running flaw-free for 30 days starting September 1. The project began in earnest on about June 1. We don't need to tell you, the project was successful, but only just, completed mere hours earlier. ...Yes, it's possible...

Customers will be using more features than LaRC is using, but largely they have more time and a larger staff. We are delighted to share our concerns but would prefer to do that in person.

3.

Cost estimate

Question:

The customer requests, but does not require, respondents to provide a cost estimate for the system described in response to this white paper. This information may include specific prices or general cost estimates for any or all hardware and software components, support licenses, set-up and installation charges, on-going maintenance, training, etc. The information will be reviewed solely by the responsible procurement officer to determine an estimated cost range for this project. As an example, if the customer receives ten cost estimates, the range will be established by the lowest and highest values. Cost Estimate information shall remain confidential until such time as it is needed.

Response:

We are grateful that (this particular customer) does not require this estimate because it would require a quotation to be made and a quotation cannot be made given the information we have at this time.

4.

Extensibility

Question:

Please describe how your system would achieve extension of functions?

Specific additional functionality might include:

a. Analytical tools used to filter and to display results from a catalog query.
b. Data mining mechanisms to construct and to execute complex catalog queries and to display the results in an easy-to-understand format.
c. Predictive maintenance indicators and troubleshooting assistance based on synthesis of system monitoring results.

Response:

There are quite a few ways in which the BigSur System® is extended to address the interests you express. The specific implementation of many of these techniques are proprietary. In each case enumerated above, more information would be very helpful in determining just how each of your suggestions of additional functionality should, technically, be implemented. We will therefore describe a number of approaches which among them may be used individually or in combination to achieve the desired results.

The most involved subject you mention is data-mining. This topic constitutes its own scientific discipline and while an in-depth discussion of data-mining is well beyond the scope of our response, we are pleased to report that since its inception, BigSur has been put to use in data-mining because it is an especially good tool to use to manage a data-mining environment.

We have affiliation with various other scientists, including leading researchers in data-mining, notably among these is Professor Sarah Graves, University of Alabama, Huntsville. In discussions with Professor Graves, she indicated that potential clients have been in contact with her with an interest in gaining her assistance in data-mining, though, she said, that collaboration has not progressed beyond the initial stage. We have furthered that dialogue and have a technical approach for addressing clients data-mining needs. In short, we would use BigSur to automate and manage her ADaM system ( http://datamining.itsc.uah.edu/adam/index.html). We have secured Professor Graves' support for a collaborative effort to meet client's data-mining needs. We should also point out that Prof. Graves is in the process of commercializing her work, so her contribution to clients efforts can fall under the "COTS" category. (sgraves@itsc.uah.edu)

In using BigSur for data-mining, algorithms are developed - typically by our customers, or through use of a system like ADaM - which detect features of interest which are then typically tracked or stored as objects themselves, used subsequently in further analytical processing, or passed to special tools for visualization. BigSur is nice for this task because it provides the automation processing environment and provides the all-important tracking meta-data so that the lineage of each object in the system is known. With BigSur, data-mining becomes integral with normal scientific processing.

These functions are at the complete discretion of BigSur users. They may be as simple a "shell script", or as complicated as a 25 year old, badly written, no-one knows how it works Fortran program. Very often, software tools such as Matlab® or Mathematica® are used. These extensions perform processing steps which create new objects. As a separate but related issue, lots of visualization tools exist and are specialized for the data-types involved, and BigSur is sensitive to which tools to use for what type. For example, many tools handle plain-text - Microsoft Word®, some display html - Netscape®, and still others manipulate images - PhotoShop®. BigSur "knows" which tools work with any given object. "Extending functions" doesn't require extending BigSur itself, it only requires entering the necessary meta-data into BigSur. "Extending data-types" is similarly accomplished as functions typically relate to specific data-types, and so they are handled in like manner. We use these meta-data to find appropriate tools which perform visualization and display of data contained within BigSur. And, favorite visualization tools for specific data-types is managed on a per-user basis so that each individual can have their own default choice known when many choices exist.

While BigSur provides some built-in room for extension with each and every component in the environment, every once in a while it would be nice to extend the meta-data to make for easy browsing of specialized components. BigSur provides a very simple, very effective mechanism for extending the schema without disturbing any existing features whatsoever. We use this capability ourselves to permit us to enter differentiating markets - we presently have the GeoDB® and MedDb® extensions for the Geospatial and Medical markets, respectively.

We understand that the fundamentals of predictive maintenance of hardware resources are based upon three criteria: time, use cycles, and known failure indications. BigSur can provide important information to aid in predictive maintenance, such as numbers of cycles, and reporting of errors. Collection of these data for analysis can be aided in several ways. For example, when our software automates the creation of magnetic media copies for distribution, the number of cycles on specific hardware can be directly tracked. When a failure or less severe error occurs, client's staff is easily able to direct the error-handling code to automatically save the pertinent information. In the case of tape writing instruments, for example, it is probably most appropriate to create an entry to manage the meta-data about that instrument as an object within BigSur. Either our built-in extension space, or new meta-data space (as outlined above) may be used to retain instrument-specific data. Subsequent analysis can be automated as just another job within the system - queued to run on a regular basis, when some threshold is reached, or by some other criteria such as operator whim. The analysis, when complete, can use BigSur mechanisms to email support staff as appropriate, or, using dialer hardware we do not (yet?) supply, the system can page staff members.

Predictive maintenance of software assets is not as straight forward but is based upon the behaviors of resource utilization and the potential need for archival of older information to keep resources free for good performance. The highest order concerns have to do with architecture and the principals of design used in developing a piece of software or software system. From the beginnings of the BigSur System®, our philosophies have been based upon the following principals and perspectives:

  • Design for performance at every juncture, and presume worst case conditions are normal. (Assume millions of database entries and thousands of concurrent users.)
  • Ours is a meta-data based system and the data is what the data is. Keep the meta-data organized so that duplication is minimized and "expensive" features are kept clear from nexus points. In other words, put "large" attributes in their own spaces, and where costly searches are likely, ensure the smallest scope of other users or data are impacted.
  • Disk space is cheap - all objects are inserted and never deleted. There should never be a need to archive data out of the environment. There should be very few times when data is updated and even fewer times when data is deleted.

As a practical matter, BigSur itself will be limited by the needs and concerns of the RDBMS that hosts it and the sufficiency of resources devoted to it, such as disk space, and CPU power. There will, generally, be no need to perform weekly, monthly or even yearly purges and only that maintenance recommended by your chosen database vendor will be required. There may at times be insufficient resources on one platform or host computer and it may therefore be desirable to move components to a second installation. This can be readily accomplished through traditional DBMS tools as provided by your DBMS vendor, or we can assist in such efforts. And, in those cases where such a move would "tear" the meta-data, our publishing software may provide an easy and direct means of populating a second installation and establishing proper meta-data links so that the meta-data remains complete and intact though the data resides on multiple installations - fully distributed meta-data is an integral feature of our system. There should be no further maintenance required for BigSur.

There is another issue related to the health and potential need for maintenance of a database, BigSur or otherwise: Referential integrity. Given that we do not know what applications will eventually be applied, and we do not know that said applications will be thoroughly written, there exists a risk that at times bad entries in the database may be made by an errant application program. Such occurrences are violations of referential integrity. Because this is a universal concern for all databases, most database systems today provide for the definition of rules, usually called "constraints," which are designed to prevent referential integrity violations. We have developed a number of referential integrity constraints for use with our database, should your RDBMS provide constraint-checking features.

And, there exist a great many tools, mostly provided by database vendors or their associated partners, which are designed to detect and potentially help repair, referential integrity violations. The best application development tools provide for the definition of rules - sometimes called "business rules", the logical equivalent of constraints - which are designed to ensure such violations never occur. (Use of our Java API will also help prevent such occurrences.) To help illustrate this issue, consider a person entered into to the system who works for Oak Ridge National Laboratory, but for whom there was never entered an affiliated organization. A query to discover if the person is with ORNL, for example, would fail to uncover the association. By teaching your applications these constraints, this maintenance concern can be reduced or eliminated altogether.

So, when it comes to predictive software maintenance of our system, this issue is primarily driven by your choice of database vendor.

5.

Customized Data-Views

Question:

2. Please describe how your system would achieve customization of data views for particular users or applications; e.g., custom portals for product customer, business office staff, maintenance personnel, department managers, and customer support specialists.

Response:

Technically speaking, data views are a technology provided by RDBMSes (Relational Database Management Systems) and implement a database query or set of queries on an underlying data-store and present to the user the results in a form that appear as if it were the underlying data-store itself. These custom views of the data are indeed used to tailor perspectives into a database for particular users or applications. Such technologies are available within most all modern RDBMSes and do not represent a component of BigSur. Creating and using views of a BigSur database is as simple or as difficult as your choice of database vendor has made it, though typically, they are all pretty simple. Where things get more interesting is in a variant called stored-procedures. The implementation of these are unique to each vendor.

Beyond the question asked, we suspect that the following additional commentary may be helpful: We ourselves are using three techniques for accessing our own installations and creating our own applications: DBMS tools provided by our RDBMS vendor(s), applications we've written in Java using our STDB Java API toolkit, and using specialized database-savvy application and development analytical tools used to filter and to display results from catalog queries. Most times, the queries are so simple that no additional help is necessary and straight SQL queries produce exactly the desired result though we recognize the security aspect of this technique is not robust. On other occasions, a script-like tool is easiest and for that we use Java-based tools from the command line, for which we have developed a security technology for which we are seeking a patent. And at still other times formal applications are desired.

Java-based programs all operate in the same fundamental way, and can be deployed in three distinct ways: command-line interfaces, direct GUI (Graphical User Interface) applications, and web-based GUIs. The advantages of Java are ubiquity, ease of programming, and the helping-hand of our Java API. Note also that in the past we have written some applications in Tcl/TK and sometimes in Perl. These applications are fine, but we have abandoned these languages for current development in favor of Java primarily because of the ubiquity of Java.

For those times when we want a sophisticated formal application, like you, we don't want to have to spend a lot of time programming or maintaining our programs. Thankfully, there is no need for us to write our own sophisticated "code-less application development" tools as these tools now abound. Among these we have used Illustra's® Object Knowledge® which is simply fantastic, and when it became unavailable, we switched to Informix's® Visionary®, equally fantastic. In both of the above cases, applications of outstanding quality are created by placing windows-like elements on a canvas and then describing what data is presented where, or which actions occur when the user manipulates something. We have ourselves created world-class applications with these tools, but unfortunately, while Visionary® is available today, we do not have a license to present our applications outside of our own offices. IBM®, which just purchased Informix Software®, has promised to continue to offer Visionary®, but has removed Visionary® from their partner program. So, we have secured an arrangement with Scopeland® Software for their Scopeland 2000® application development and deployment tools and our license permits us to deploy our applications on the public web. We have the tools in-house and have begun development using them. While not as sophisticated as Visionary®, Scopeland 2000® is much less expensive, and is doing the job quite well. Scopeland 2000Ò is designed for use by people with no programming skills whatsoever. We began developing an application the moment the software was installed and development is indeed quite rapid. We anticipate a couple of weeks spent on the learning curve, followed by a couple of weeks spent per application - for one "programmer". We will deploy these applications on one of our internet accessible servers.

We have established ourselves as a partner of Scopeland Software and will be offering their products and services as an optional part of our proposal to you.

In addition to these vendors, we have become aware of no less than four other vendors who serve the codeless-application-development market. Some of them have capabilities well beyond Scopeland, while others supply much less, yet all of them are complementary to our offerings and all of them greatly simplify the creation and deployment of suitable applications via web-servers. As our dialogues with them and evaluation of their products are not as far along, we are hesitant to cite them by name. However, we would gladly do so, if asked.

 


Set One | Set Two | Set Three | Set Four | Set Five | Set Six | All Topics

 

 
Feedback
Contact Us

website contact: Webmistress

Science Tools > Top Level