Why the multiple ECM Repository/Silo model is not a good idea

by Frank 15. November 2016 06:00

“43 Reasons why Managing Records in-Place may not be good enough”

Enterprise Content Management is a moving target, constantly evolving with new challenges and new paradigms. For example, how do we filter out only relevant information from social media? How do we avoid capturing personal data and being culpable under privacy laws? How do we capture all emails containing sexism, racism and bullying without being guilty of an invasion of privacy of the individual? How do we meet all of our compliance obligations when our staff are spread across multiple states/counties/provinces and multiple countries with different legislation and compliance requirements? All weighty challenges for the modern Knowledge Manager or CIO.

Another interesting challenge for Knowledge Managers and CIOs is the newer document management paradigm of being asked to manage all content without a single central repository. That is, to be responsible for all content across a myriad of locations controlled by a myriad of applications and a myriad of departments/organizations and people. Back when I was an employee and not an employer, my tough (ex-military) manager in Blue Bell, PA would just bang his fist on his desk and say, “Goddam Frank, just do it!” That was always a signal for me to get creative.

However, try as I may, I am finding it nigh on impossible to get creative enough to work out how I could effectively and reliably manage all content across an enterprise without a single central repository.

In multiple-repository systems we find multiple document stores; local files, network file shares, local data bases, multiple file servers, multiple copies of SharePoint and multiple Cloud repositories like Dropbox, Box, iCloud, Google Cloud Storage and other hosted document storage. The CIO may proudly claim to manage multiple information silos but what he or she really has is a laissez faire document management ecosystem that may well be centrally monitored (hopefully) but is most certainly not centrally managed.

In the multiple silo model the documents in our multiple locations are ‘managed’ by multiple people and multiple applications (e.g., SharePoint, Google Docs, etc.). We may have implemented another layer of software above all these diverse applications trying to keep up with what is happening but If I am just ‘watching’ then I don’t have an inviolate copy and I don’t have any control over what happens to the document. I am unable to enforce any standards. There is no ‘standard’ central control over versioning or retention and no control over the document life cycle or chain of evidence.

For example, you wouldn’t know if the document had since been moved to a different location that you are not monitoring. You wouldn’t know if it had been deleted. You wouldn’t know its relationship to other documents and processes in other silos. You wouldn’t know its context in your enterprise and therefore you wouldn’t know how relevant this document was. The important distinction is that under the multiple silo model you are ‘watching’ not managing; other software is managing the life-cycle and disposition of the document.

All you really know is that at a certain point in time a document existed and what its properties were at that time (e.g., historical ‘natural’ Metadata such as original filename, author, date created, etc.). However, you have no contextual Metadata, no transactional Metadata, no common indexing and no common Business Classification System. In this case, you don’t have a document management system, you have a laissez faire document management ecosystem, an assortment of independently ‘managed’ information silos. Most importantly, you are not able to link documents to business processes that transcend organizational structures and silos.

Sure, SharePoint and Cloud silos make collaboration easier but at what cost? What can’t we do with this multi-silo ecosystem? Why doesn’t this solution meet the best-practice objectives of a document management system? What are the major areas where it falls short? How does the proliferation of multiple silos and content repositories affect us? What are our risks? Here is my assessment of the major shortfalls of this paradigm.

 We are unable to:

1.    extract the critical insights that enterprise information should provide

2.    define all the relationships that link documents to enterprise business processes

3.    find the right information at the right time

4.    provide a single access point for all content

5.    Implement an effective, consistent enterprise-wide document security system

6.    effectively protect against natural or man-made disasters

7.    produce evidence-standard documents

8.    minimize document handling costs

9.    guarantee the integrity of a document

10.guarantee that a document is in fact the most recent version

11.guarantee that a document is not an older copy

12.minimize duplicate and redundant information

13.meet critical compliance targets like Sarbanes-Oxley Act (SOX) and the HIPAA

14.create secure, searchable archives for digital content

15.effectively secure all documents against loss

16.implement common enterprise version control

17.facilitate enterprise collaboration

18.Improve timeliness

19.manage enterprise document security and control

20.manage smaller and more reliable backups

21.achieve the lowest possible document management and archiving costs

22.deliver the best possible knowledge management access and search

23.guarantee consistent content

24.optimize management and executive time

25.standardize the types of documents and other content can be created within an organization.

26.define common use template to use for each type of document.

27.standardize the Metadata required for each type of document.

28.standardize where to store a document at each stage of its life cycle.

29.control access to a document at each stage of its life cycle.

30.move documents within the organization as team members contribute to the documents' creation, review, approval, publication, and disposition.

31.implement a common set of policies that apply to documents so that document-related actions are audited, documents are retained or disposed of properly, and content that is important to the organization is protected.

32.manage when and if a document has to be converted from one format to another as it moves through the stages of its life cycle.

33.guarantee that all documents are treated as corporate records, that common retention policies are applied determining which documents must be retained according to legal requirements and corporate guidelines.

34.guarantee enterprise-wide Regulatory compliance

35.produce an enterprise-wide audit trail

36.share information across departmental and/or silo boundaries

37.centrally manage the security access to documents/information across different areas of the organization.

38.consistently classify documents as each repository may be used by a different department and be classified differently.  

39.identify duplicates based on document name.

40.easily find things based on metadata, as it wouldn’t be common across repositories.

41.control access via AD single sign on

42.access all enterprise documents using a single license.

          43.centrally audit access and changes to metadata.

What are your risks?  Your risks are huge!

 

 

 

 

 

How to simplify electronic document and email management

by Frank 17. September 2014 06:00

I have written about this topic many times in the past (see links at the end of this post) but the lesson is always the same. There are two key rules:

1.     If your system relies on people being 100% consistent and reliable it won’t work; and

2.     If you system places an additional workload on already busy workers it won’t work.

The message is, if you simplify and automate your system you give it the best possible chance of working.

If your system works as automatically as possible and doesn’t require much effort from your workforce then it has the best possible chance of being successful.

With today’s technology and tools there is simply no need to burden your workforce with capture and classification tasks. Do you still see people still using typewriters, rotary phones or Morse code? No you don’t because there is much better technology available. So why do you persist with an old, outdated and unsuccessful model? Why do you ask your staff to manually capture and classify electronic documents and emails when there are much better, much faster, much more consistent and much more reliable ways to achieve a better result? It is after all 2014, not 1914; we all use computers and smart phones now, not typewriters, wind-up rotary phones and Morse code.

Emails are managed by email servers, (yes, even Google). Email servers allow plug-ins and add-ons and are ‘open’ so you can automatically monitor and capture incoming and outgoing emails.

Electronic documents are always saved somewhere, for example on your shared drives or directly into your DMS. As such they can be captured and interrogated programmatically.

It is entirely possible to ‘parse’ any electronic document or email and its associated attributes and Metadata and make consistent decisions about whether or not to capture it and how to classify it when captured. It isn’t rocket science any more, it is just analysis, design and programming. We can go even further and determine who should be notified and what action(s) need to be initiated in response to each new email or electronic document.  

We can easily implement an end-to-end business process whereby every electronic document and email is managed from creation to destruction and we can do this with minimal human involvement. Where human involvement is required, for example making a decision or deciding upon an appropriate response, we can also automate and manage the business processes required and simply ‘present’ staff with all the required information when required.

Isn’t this was the Knowledge Management revolution was supposed to be about?

“A system that provides the user with the explicit information required, in exactly the form required at precisely the time the user needs it.”

The new model is all about automation and processing at the server rather than at the user’s workstation; a fully automatic, server-centric paradigm. A system that is all about the ‘Push’ rather than the ‘Pull’ model. A model whereby the computer services the end user, where the end user is not a slave to the computer.

We could also call it management by exception. “Please only give me what I need to see when I need to see it.”

None of the above is new or revolutionary thinking, it is all just common sense. None of the above requires yet-to-be invented technology or products, it only requires existing and proven technology and products.

The fully-automatic, server-centric approach should be the default choice and it should be a no-brainer for any organization that needs to implement an email and document management regime. Unfortunately, too often it isn’t.

If you have the responsibility of rolling out an email and document management system and the fully-automatic, server-centric approach isn’t on your agenda then your boss should be asking you why not.

References:

White papers

Posts

How to clean up your shared drives, Frank’s approach

by Frank 22. August 2014 06:00

In my time in this business (enterprise content management, records management, document management, etc.) I have been asked to help with a ‘shared drive problem’ more times than I can remember. This particular issue is analogous with the paperless office problem. Thirty years ago when I started my company I naively thought that both problems would be long gone by now but they are not.

I still get requests for purely physical records management solutions and I still get requests to assist customers in sorting out their shared drives problems.

The tools and procedures to solve both problems have been around for a long time but for whatever reason (I suspect lack of management focus) the problems still persist and could be described as systemic across most industry segments.

Yes, I know that you can implement an electronic document and records management system (we have one called RecFind 6) and take away the need for shared drives and physical records management systems completely but most organizations don’t and most organizations still struggle with shared drives and physical records. This post addresses the reality.

Unfortunately, the most important ingredient in any solution is ‘ownership’ and that is as hard to find as it ever was. Someone with authority, or someone who is prepared to assume authority, needs to take ownership of the problem in a benevolent dictator way and just steam-roll a solution through the enterprise. It isn’t solvable by committees and it requires a committed, driven person to make it happen. These kind of people are in short supply so if you don’t have one, bring one in.

In a nutshell there are three basic problems apart from ownership of the problem.

1.     How to delete all redundant information;

2.     How to structure the ‘new’ shared drives; and

3.     How to make the new system work to most people’s satisfaction.

Deleting redundant Information

Rule number one is don’t ever ask staff to delete the information they regard as redundant. It will never happen. Instead, tell staff that you will delete all documents in your shared drives with a created or last updated date greater than a nominated date (say one-year into the past) unless they tell you specifically which ‘older’ documents they need to retain. Just saying “all of them” is not an acceptable response. Give staff advance notice of a month and then delete everything that has not been nominated as important enough to retain.  Of course, take a backup of everything before you delete, just in case. This is tough love, not stupidity.

Structuring the new shared drives

If your records manager insists on using your already overly complex, hierarchical corporate classification scheme or taxonomy as the model for the new shared drive structure politely ask them to look for another job. Do you want this to work or not?

Records managers and archivists and librarians (and scientists) understand and love complex classification systems. However, end users don’t understand them, don’t like them and won’t use them. End users have no wish to become part-time records managers, they have their own work to do thank you.

By all means make the new structure a subset of the classification system, major headings only and no more than two levels if possible. If it takes longer than a few seconds to decide where to save something or to find something then it is too complex. If three people save the same document in three different places then it is too complex. If a senior manager can’t find something instantly then it is too complex. The staff aren’t to blame, you are.

I have written about this issue previously and you can reference a white paper at this link, “Do you really need a Taxonomy?”

The shared drives aren’t where we classify documents, it is where we make it as easy and as fast as possible to save, retrieve and work on documents; no more, no less. Proper classification (if I can use that term) happens later when you use intelligent software to automatically capture, analyse and store documents in your document management system.

Please note, shared drives are not a document management system and a document management system should never just be a copy of your shared drives. They have different jobs to do.

Making the new system work

Let’s fall back on one of the oldest acronyms in business, KISS, “Keep It Simple Stupid!” Simple is good and elegant, complex is bad and unfathomable.

Testing is a good example of where the KISS principle must be applied. Asking all staff to participate in the testing process may be diplomatic but it is also suicidal. You need to select your testers. You need to pick a small number of smart people from all levels of your organization. Don’t ask for volunteers, you will get the wrong people applying. Do you want participants who are committed to the system working, or those who are committed to it failing? Do you want this to succeed or not?

If I am pressed for time I use what I call the straight-line-method. Imagine all staff in a straight line from the most junior to the most senior. Select from both ends, the most junior and the most senior. Chances are that if the system works for this subset that it will also work for all the staff in between.

Make it clear to all that the shared drives are not your document management system. The shared drives are there for ease of access and to work on documents. The document management system has business rules to ensure that you have inviolate copies of important documents plus all relevant contextual information. The document management system is where you apply business rules and workflow. The document management system is all about business process management and compliance. The shared drives and the document management system are related and integrated but they have different jobs to do.

We have shared drives so staff don’t work on documents on ‘private’ drives, inaccessible and invisible to others. We provide a shared drive resource so staff can collaborate and share information and easily work on documents. We have shared drives so that when someone leaves we still have all their documents and work-in-process.

Please do all the complex processes required in your document management system using intelligent software, automate as much as possible. Productivity gains come about when you take work off staff, not when you load them up with more work. Give your staff as much time as possible so they can use their expertise to do the core job they were hired for.

If you don’t force extra work on your staff and if you make it as easy and as fast as possible to use the shared drives then your system will work. Do the opposite and I guarantee it will not work.

Using barcodes to raise productivity and lower costs in Records Management processes

by Frank 6. August 2014 06:00

Did you know that in the spring of 1969 the first true bar code systems were installed? One went into a General Motors plant in Pontiac, Michigan, where it was used to monitor the production and distribution of automobile axle units. The other went into a distribution facility run by General Trading Company in Carlsbad, New Jersey, to help direct shipments to the proper loading-bay doors.

Did you also know that the very first product to be sold with a barcode and scanner was a single packet of chewing gum at a Marsh supermarket in Troy, Ohio on June 26, 1974?

Both these interesting facts came from an excellent article on the history of barcodes by Tony Seideman. Please see this link.

The overall advantages and benefits of barcodes are well known; speed, accuracy, ease of implementation and cost-effectiveness.

In a nutshell, barcodes are cheap to produce, easy to implement and easy to read. They are infinitely better than a human keying in information. Barcodes are reliable and they just work.

Modern supermarkets simply couldn’t function without barcodes on products and barcode readers at checkouts.

Most well-run records management facilities also use barcodes to great advantage to track file-folders and boxes, run audits and speed up the entering of information. Most offsite records storage facilities use barcodes to track boxes on shelves. It is what we call a “no brainer.”

However, despite the obvious benefits, especially the cost benefits, many organizations today still manage physical assets bereft of barcodes. You may well ask “why?” and so do I. Given the low cost of both barcodes and barcode readers and the well-proven technology, I honestly can’t think of any reason for not using barcoding technology to manage physical assets like file-folders and archive boxes. It just doesn’t make any sense whatever to me. It is analogous to running ten miles to deliver a message rather than just phoning or texting. How many messages a day can you deliver by running and how many can you deliver a day by phoning or texting?

Why ask staff to write down file-folder numbers or enter them on a keyboard when you can ‘wand’ or 'scan' them much more accurately and infinitely faster using a barcode reader? Why put up with processing 20 file movements a day by hand when you can easily process 200 a day using a barcode reader?

If you have 30 file-folders on your desk that you have to process why would you do it manually by keying in each file number (and making mistakes) over 30 minutes when you could process the same number of file-folders in 30 seconds using a fixed barcode reader (and not making any keying mistakes)?

When you have 500 file-folders to add to archive boxes provided by your offsite storage provider why would you take hours to do it laboriously with lists and the keyboard when you could do it in minutes using a barcode reader? Simply use your portable barcode reader to read the box barcode then read each file-folder barcode number as you add it to the box and then read the box number again when finished to complete the transaction. What could be faster or simpler?

So, what do you need to convert your slow and error-prone manual-entry records management processes to fast and accurate barcode-enabled processes?

1.       A records management software package that supports barcodes (I don’t know of any modern RM system that doesn’t)

2.       *A supply of pre-printed barcodes (or you can print them out of your records management software package)

3.       Some fixed or wedge barcode readers (expect to pay $150 to $250 each)

4.       One or more portable barcode readers (expect to pay $1,000 to $2,000 including cables, battery chargers, etc.)

*A word on barcode labels. It pays to make them as durable as possible. This usually means laminating them as un-laminated barcodes produced on a laser printer tend to have a short life expectancy. The easiest way to obtain high quality, laminated barcode labels is to order them from a specialist print house. This way you can specify exactly what you need in terms of format and size and be assured of a long life and reliability. Nothing frustrates more than a worn barcode that doesn’t read properly.

Of course someone has to stick the barcode label on the file-folders and then tell the computer system (i.e., file-folder number AB/2003/00067 is now barcode number 1000049). You have a choice of how to do this. If you don’t have too many file-folders you can bite the bullet and add them all as a special project. Or, you can decide just to add them to every new file-folder created and to add barcodes to existing file-folders when they cross your desk. It is your decision based on volume and resources. However, you need to invest the effort to reap the benefits.

Then if you really want to benefit you will assign a different class of barcode to ‘locations’. That is, offices, shelves, rooms, etc., and even people. This is so you can do an audit on a regular basis using your portable barcode reader. Wouldn’t it be nice to know where everything is and even, where some things aren’t?

Finally, assign yet another set of barcodes to your archive boxes so it is as easy and as fast as possible to move file-folders into and out of archive boxes.

The above describes just the simplest application of barcodes but even so, the benefits and cost savings are significant. The more creative of you will comes up with many more ways to make barcodes pay big dividends. We have one customer for example, that automatically allocates barcodes to emails in Outlook to make them easier to monitor and track both electronically and physically. See this link:

Barcodes are simple to use, low cost and well-proven, ‘risk-free’ technology. The effective use of barcodes and barcode readers can remove drudgery, lower costs and massively improve productivity.

If you aren’t using barcodes your boss should be asking you “why not?”

Document Imaging, Forms Processing & Workflow – A Guide

by Frank 28. July 2014 06:00

Document imaging (scanning) has been a part of most business processing since the early 1980s. We for example, produced our first document imaging enabled version of RecFind in 1987. So it isn’t new technology and it is now low risk, tried and proven technology.

Even in this age of electronic documents most of us still receive and have to read, analyse and process mountains of paper.

I don’t know of any organization that doesn’t use some form of document imaging to help process paper documents. Conversely, I know of very few organizations that take full advantage of document imaging to gain maximum value from document imaging.

For example, just scanning a document as a TIFF file and then storing it on a hard drive somewhere is almost a waste of time. Sure, you can then get rid of the original paper (but most don’t) but you have added very little value to your business.

Similarly, capturing a paper document without contextual information (Metadata) is not smart because you have the document but none of the important transactional information. Even converting a TIFF document to a PDF isn’t smart unless you first OCR (Optical Character Recognition) it to release the important text ‘hidden’ in the TIFF file.

I would go even further and say that if you are not taking the opportunity to ‘read’ and ‘capture’ key information from the scanned document during the scanning process (Forms Processing) then you aren’t adding anywhere near as much value as you could.

And finally, if you aren’t automatically initiating workflow as the document is stored in your database then you are criminally missing an opportunity to automate and speed up your internal business processes.

To give it a rating scale, just scanning and storing TIFF files is a 2 out of 10. If this is your score you should be ashamed to be taking a pay packet. If you are scanning, capturing contextual data, OCRing, Forms Processing, storing as a text-searchable PDF and initiating workflow then you get a 10 out of 10 and you should be asking your boss for a substantial raise and a promotion.

How do you rate on a scale of 0 to 10? How satisfied is your boss with your work? Are you in line for a raise and a promotion?

Back in the 1980s the technology was high-risk, expensive and proprietary and few organizations could afford the substantial investment required to scan and process information with workflow.

Today the technology is low cost and ubiquitous. There is no excuse for not taking full advantage of document imaging functionality.

So, where do you start?

As always, you should begin with a paper-flow analysis. Someone needs to do an inventory of all the paper you receive and produce and then document the business processes it becomes part of.

For every piece of paper you produce you should be asking “why?” Why are you producing paper when you could be producing an electronic document or an electronic form?

In addition, why are you producing multiple copies? Why are you filing multiple copies? What do your staff actually do with the paper? What happens to the paper when it has been processed? Why is it sitting in boxes in expensive off-site storage? Why are you paying to rent space for that paper month after month after month? Is there anything stored there that could cause you pain in any future legal action?

And most importantly, what paper can you dispose of?

For the paper you receive you need to work out what is essential and what can be discarded. You should also talk to your customers, partners and suppliers and investigate if paper can be replaced by electronic documents or electronic forms. Weed out the non-essential and replace whatever you can with electronic documents and electronic forms. For example, provide your customers, partners and suppliers with Adobe electronic forms to complete, sign and return or provide electronic forms on your website for them to complete and submit.

Paper is the enemy, don’t let it win!

Once you have culled all the paper you can, you then need to work out how to process the remaining paper in the most efficient and effective manner possible and that always ends up as a Business Process Management (BPM) exercise. The objectives are speed, accuracy, productivity and automation.

Don’t do anything manually if you can possibly automate it. This isn’t 30 years ago when staff were relatively cheap and computers were very expensive. This is now when staff are very expensive and computers are very cheap (or should I say low-cost?).

If you have to process paper the only time it should be handled is when it is taken from the envelope and fed into a document scanner. After that, everything should be automated and electronic. Yes, your records management department will dutifully want to file paper in file folders and archive boxes but even that may not be necessary.  Don’t accept the mystical term ‘compliance’ as a reason for storing paper until you really do understand the compliance legislation that applies to your business. In most cases, electronic copies, given certain safeguards, are acceptable.

I am willing to bet that your records manager will be operating off a retention schedule that is old, out-of-date, modified from another schedule, copied, modified again and ‘made-to-fit’ your needs. It won’t be his/her fault because I can probably guarantee that no budget was allocated to update the retention schedule on an ongoing basis. I am also willing to bet that no one has a copy of all of the current compliance rules that apply to your business.

In my experience, ninety-percent plus of the retention schedules in use are old, out-of-date and inappropriate for the business processes they are being applied to. Most are also way too complicated and crying out for simplification. Bad retention schedules (and bad retention practices – are you really destroying everything as soon as you are allowed?) are the main reason you are wasting thousands or millions of dollars a year on redundant offsite storage.

Do your research and save a fortune! Yes, records are very important and do deserve your attention because if they don’t get your attention you will waste a lot of money and sooner or later you will be penalised for holding information you could have legally destroyed a long time ago. A good records practice is an essential part of any corporate risk management regime. Ignore this advice at your peril.

Obviously, processing records efficiently requires software. You need a software package that can:

  1. Scan, OCR and Forms Process paper documents.
  2. Capture and store scanned images and associated Metadata plus any other kind of electronic document.
  3. Define and execute workflow.
  4. Provide search and inquiry capabilities
  5. Provide reporting capabilities.
  6. Audit all transactions.

The above is obviously a ‘short-list’ of the functionality required but you get the idea. There must be at least several hundred proven software packages in the world that have the functionality required. Look under the categories of:

  1. Enterprise Content Management (ECM, ECMS)
  2. Records Management (RM, RMS)
  3. Records and Document Management
  4. Document Management (DM, DMS)
  5. Electronic Document and Records Management (EDRMS)
  6. Business Process Management (BPM)

You need to define your business processing requirements beginning with the paper flow analysis mentioned earlier. Then convert your business processing requirements into workflows in your software package. Design any electronic forms required and where possible, re-design input paper forms to facilitate forms processing. Draw up procedures, train your staff and then test and go live.

The above paragraph is obviously a little short on detail but I am not writing a “how-to” textbook, just a simple guide. If you don’t have the necessary expertise then hire a suitably qualified and experienced consultant (someone who has done it before many times) and get productive.

Or, you can just put it off again and hope that you don’t get caught.

 

Are you still struggling with physical records management, with paper?

by Frank 16. July 2014 00:01

 

Are you still struggling with physical records management, with paper?

We produced our first computerised records management system in 1984 (when our company was called GMB) and it was called DocFind. It was marketed by the Burroughs Corporation initially to about 100 clients and then we stared marketing DocFind direct and sold it to about another 2,000 clients.

Every one of those clients wanted DocFind just to manage physical records, paper, file folders and archive boxes. There was little or no demand for document imaging and workflow and the term electronic document management had yet to be invented. Office automation was in its infancy. We for example, wrote our letters on an Apple IIe using a word processor called WordStar running under CP/M.

In 1986 we released RecFind, a major remake of the DocFind product. This product was initially marketed by ourselves and NEC and it too focussed just on managing physical records.

However, even in 1986 we knew we had a bigger job to do with the general acceptance of document scanners and workflow so we added imaging and workflow to our product and starting trying to convince our customers and prospective customers to reduce the size of their paper mountain and even to start planning for a ‘Paperless Office’.

In the late 1980s and early 1990s I delivered numerous papers extolling the value of the paperless office and worked hard to convince my customers to make the move to Electronic Document and Records Management (EDRMS).

In the mid-1990s the industry discovered ‘Knowledge Management’ (KMS) and industry consultants lost interest in EDRMS and instead heavily promoted the virtues and benefits of KMS, whatever it was. Maybe this was the time organizations lost interest in eradicating paper as senior IT staff and consultants moved on to more interesting projects like KMS.

In 1995 I delivered my first paper on a totally integrated information management system or what I called at the time the ‘It Does Everything Application’ (IDEA). In 1995 I truly thought the age of physical records management was almost over and that the western world at least would move to fully-automated, paperless processes.

How wrong I was 19 years ago.

Today, despite the advanced functionality of our RecFind 6 Product Suite, almost all of my customers still manage physical records with RecFind 6. At least half of the inquiries that come in via our website are for systems to manage physical records.

There is more paper in the world today than there has ever been and organizations all over the world still struggle with managing paper, vast amounts of paper.

Luckily for us, we never succumbed to the temptation to remove the paper handling features from our products. Instead, we added to them with each subsequent release and redesign/rewrite of RecFind. We had to provide upwards compatibility for our clients as they still managed mountains of paper both onsite and offsite.

Being a little older and wiser now I am never again going to predict the paperless office. I will provide advanced physical records management functionality for my clients as long as they require it.

I haven’t given up the fight but my job is to address the real needs of my customers and they tell me and keep telling me that they need to manager paper, mainly file folders full of paper and archive boxes full of file folders. They need to manage paper onsite in shelving and offsite in warehouses with millions of boxes and we do it all.

We manage paper from creation to destruction and throughout the whole lifecycle. We apply retention schedules and classification systems and we track anything and everything with barcodes and barcode readers. We have enhanced our products to cater for every need and we are now probably responsible for millions of tonnes of paper all over the world.

I still hope for a paperless world but I very much doubt that I am going to see it in my lifetime.

So, if you are still struggling with how to best manage all your physical records please don’t despair, you are most certainly not alone! 

  

What is the future of RecFind? - The Product Road Map

by Frank 19. May 2014 06:00

First a little history. We began in 1984 with our first document management application called DocFind marketed by the then Burroughs Corporation (now called Unisys). In June 1986 we sold the first version of RecFind, a fully-featured electronic records management system and a vast improvement on the DocFind product. Then we progressively added document imaging then electronic document management and workflow and then with RecFind 6 a brand new paradigm and an amalgam of all previous functionality; an Information management system able to run multiple applications concurrently with a complete set of enterprise content management functionality. RecFind 6 is the eighth completely new iteration of the iconic RecFind brand.

RecFind 6 was and is unique in our industry because it was designed to be what was previously called a Rapid Application Development system (RAD) but unlike previous examples, we provided the high level toolset so new applications could be inexpensively ‘configured’ (by using the DRM) not expensively programmed and new application tables and fields easily populated using Xchange. It immediately provided every customer with the ability to change almost anything they needed changed without needing to deal with the vendor (us).  Each customer had the same tools we used to configure multiple applications within a single copy of RecFind 6. RecFind 6 was the first ECM product to truly empower the customer and to release them from the expensive and time consuming process of having to negotiate with the vendor to “make changes and get things done.”

In essence, the future of the RecFind brand can be summarised as more of the same but as an even easier to use and more powerful product. Architecturally, we are moving away from the fat-client model (in our case based on the .NET smart-client paradigm) to the zero-footprint, thin-client model to reduce installation and maintenance costs and to support far more operating system platforms than just Microsoft Windows. The new version 2.6 web-client for instance happily runs on my iPad within the Safari browser and provides me with all the information I need on my customers when I travel or work from home (we use RecFind 6 as our Customer Relationship Management system or CRM). I no longer need a PC at home and nor do I need to carry a heavy laptop through airports.

One of my goals for the remainder of 2014 and 2015 following is to convince my customer base to move to the RecFind 6 web-client from the standard .NET smart-client. This is because the web-client provides tangible, measurable cost benefits and will be the basis for a host of new features as we gradually deprecate the .NET smart-client and expand the functionality of the web-client. We do not believe there is a future for the fat/smart-client paradigm; it has seen its day. Customers are rightfully demanding a zero footprint and the support of an extensive range of operating environments and devices including mobile devices such as smartphones and tablets. Our web-client provides the functionality, mobile device support and convenience they are demanding.

Of course the back-end of the product, the image and data repository, also comes in for major upgrades and improvements. We are sticking with MS SQL Server as our database but will incorporate a host of new features and improvements to better facilitate the handling of ‘big data’. We will continue to research and make improvements to the way we capture, store and retrieve data and because our customer’s databases are now so large (measured in hundreds of Gigabytes), we are making it easier and faster to both backup and audit the repository. The objectives as always are scalability, speed, security and robustness.

We are also adding new functionality to allow the customer to bypass our standard user interface (e.g., the .NET smart-client or web-client) and create their own user interface or presentation layer. The objective is to make it as easy as possible for the customer to create tailored interfaces for each operating unit within their organization. A simple way to think of this functionality is to imagine a single high level tool that lets you quickly and easily create your own screens and dashboards and program to our SDK.

On the add-in product front we will continue to invest in our add-in products such as the Button, the MINI API, the SDK, GEM, RecCapture, the High Speed Scanning Module and the SharePoint Integration Module. Even though the base product RecFind 6 has a full complement of enterprise content management functionality these add-on products provide options requested by our customers. They are generally a way to do things faster and more automatically.

We will continue to provide two approaches for document management; the end-user paradigm (RecFind 6 plus the Button) and the fully automatic capture and classification paradigm (RecFind 6 plus GEM and RecCapture). As has been the case, we also fully expect a lot of our customers to combine both paradigms in a hybrid solution.

The major architectural change is away from the .NET smart-client (fat-client) paradigm to the browser-based thin-client or web-client paradigm. We see this as the future for all application software, unconstrained by the strictures of proprietary operating systems like Microsoft Windows.

As always, our approach, our credo, is that we do all the hard work so you don’t have to. We provide the feature rich, scalable and robust image and data repository and we also provide all of the high level tools so you can configure your applications that access our repository. We also continue to invest in supporting and enhancing all of our products making sure that they have the feature set you require and run in the operating environments you require them to. We invest in the ongoing development of our products to protect your investment in our products. This is our responsibility and our contribution to our ongoing partnership.

 

Technology Trends for 2014 – A developer’s perspective

by Frank 7. January 2014 06:00

I run a software company called the Knowledgeone Corporation and we produce enterprise content management software for government and business. Because it takes so long to design, build and test a new product or even a new version, we have to try and predict where the market will be in one or two years and then try to make sure our product RecFind 6 ‘fits-in’ with future requirements.

Years ago it was much easier because we were sure Windows would be the dominant factor and mostly we had to worry about compatibility with the next version of Windows and Microsoft Office. Apple however, changed the game with first the iPhone and then the iPad.

We now need to be aware of a much wider range of devices and operating systems; smart phones and tablets in particular. Three years ago we decided to design in compatibility for iOS and Android and we also decided to ignore Blackberry; so far, a wise move.

However, the prediction business is getting harder because the game is changing faster and probably faster than we can change our software (a major application).

I was just reading about CES 2014 on ZDNet and the major technologies previewed and displayed there. Most are carry overs from 2013 and I haven’t noted anything really new but even so, the question is which of these major trends will become major players during 2014 and 2015 (our design, develop and test window for the next major release of RecFind 6)?

1.     Wearables

2.     The Internet of Things

3.     Contextual Computing (or Predictive Computing)

4.     Consumerization of business tech

5.     3D printing

6.     Big Data

7.     The Cloud

Larry Dignan, Editor in Chief of ZDNet, wrote an excellent summary of things to think about for 2014, see this link:

Larry sees China and emerging Chinese companies as major players outside of China in 2014 but I think the Europeans and Americans will resist until well into 2015 or later. Coming on the heels of the Global Financial Crisis of 2008 their governments won’t take kindly to having their local high tech industries swamped by Chinese giants. He also talks about the fate of Windows 8 and the direction of the PC market and this is our major concern.

The PC market has been shrinking and even though Microsoft is still the major player by far a lot depends upon the acceptance of Windows 8 as the default operating system. Personally I saw the Windows 8 Metro interface as clumsy and as change for changes’ sake.

I really don’t understand Microsoft’s agenda. Why try to force a major change like this on consumers and businesses just when everyone is happy with Windows 7 and we have all almost forgotten Vista. Windows 8 isn’t an improvement over Windows 7 just as Office 2013 isn’t an improvement over Office 2010. Both are just different and in my opinion, less intuitive and more difficult to use.

Try as I might, I cannot see any benefits to anyone in moving from Windows 7 to Windows 8 and in moving from Office 2010 to Office 2013. The only organization benefiting would be Microsoft and at the cost of big disruptions to its loyal customers.

Surely this isn’t a wise thing to do in an era of falling PC sales? Why exacerbate the problem?

Smart phones and tablets are real and growing in importance. Android and iOS are the two most important ‘new’ operating systems to support and most importantly for us, browsers are the application carriers of the future. No software vendor has the resources to support all the manifestations of Windows, Linux, Android, iOS, etc., in ‘native’ form but all operating systems support browsers. Browsers have become what Windows was ten years ago. That is, a way to reach most of the market with a single set of source code.

We lived through the early days of DOS, UNIX, Windows and the AS/400 and at one time had about fifteen different sets of source code for RecFind. No vendor wants to go back to those bad old days. When the world settled on Windows it meant that most of us could massively simplify our development regime and revert to a single set of source code to reach ninety-percent of the market. In the early days, Windows was our entry point to the world. Today it is browsers.

Of course not all browsers are equal and there is extra work to do to support different operating systems, especially sand-boxed ones like iOS but, we are still running ninety five percent common source and five-percent variations so it is eminently manageable.

Does Microsoft realize that many developers like us now target browsers as our main application carriers and not Windows? Does it also realize that the Windows 8 Metro interface was the catalyst that pushed many more developers along this same path?

Let’s hope that the new CEO of Microsoft cares more about his customers than the previous one did. If not, 2014 won’t just be the post-PC era, it will also be the beginning of the post-Microsoft era.

Is this Microsoft’s worst mistake ever?

by Frank 30. November 2013 06:00

I run a software company called the Knowledgeone Corporation that has been developing application solutions for the Microsoft Windows platform since the very first release of Windows. As always, our latest product offering RecFind 6 version 2.6 has to be tested and certified against the latest release of windows. In this case that means Windows 8.1.

Like most organizations, we waited for the Windows 8.1 release before upgrading our workstations from Windows 7. The only exceptions were our developers workstations because we bought them new PCs with Windows 8 pre-installed.

We are now testing the final builds of RecFind 6 version 2.6 and have found a major problem. The problem is that Microsoft in its infinite wisdom has decided that you can’t install Windows 8.1 over a Windows 7 system and retain your already installed applications.

The only solution is to install Windows 8 first and then upgrade Windows 8 to Windows 8.1. However, if you are running Windows 7 Enterprise this won’t work either and you will be told that you will have reinstall all of your applications.

I am struggling to understand Microsoft’s logic.

Surely Microsoft wants all its customers to upgrade to Windows 8.1? If so, why has it ‘engineered’ the Windows 8.1 upgrade so customers will be discouraged from using it? Does anyone at Microsoft understand how much work and pain is involved in re-installing all your applications?

No, I am not kidding. If you have a PC or many PCs with Windows 7 installed you are going to have to install Windows 8 first in order to maintain all of your currently installed applications. Then, after spending many hours installing Windows 8 (it is not a trivial process) spend more precious time installing Windows 8.1. Microsoft has ensured that you cannot go direct from Windows 7 to Windows 8.1.

Of course, if you are unlucky, you could be living in a country where Microsoft has blocked the downloading of Windows 8, like Australia. Now you are between a rock and a hard place. Microsoft won’t let you install Windows 8 and if you install Windows 8.1 you face days or weeks of frustrating effort trying to re-install all of your existing applications.

 

Here are some quotes from Microsoft:

“You can decide what you want to keep on your PC. You won't be able to keep programs and settings when you upgrade. Be sure to locate your original program installation discs or purchase confirmation emails if you bought programs online. You'll need these to reinstall your programs after you upgrade to Windows 8.1—this includes, for example, Microsoft Office, Apache OpenOffice, and Adobe programs. It's also a good idea to back up your files at this time, too.”

If you're running Windows 7, Windows Vista, or Windows XP, all of your apps will need to be reinstalled using the original installation discs, or purchase confirmation emails if you bought the apps online.”

If the management at Microsoft wanted to ensure the failure of Windows 8.1 they couldn’t have come up with a better plan than the one they have implemented. By making Windows 8.1 so difficult to install they have ensured that its customers will stick with the tried and proven Windows 7 for as long as possible.

Can anyone at Microsoft explain why they thought this was a good idea?

Is the IT industry faltering because we have all just lost interest?

by Frank 15. April 2013 06:00

I have just read another IDC industry reporting talking about how PC sales have plunged 14 percent in the first three months of 2013. The report goes on to show that this is a worldwide trend, not just in the USA or Asia Pacific. Europe for example, was the worst with a 16 percent decline.

I also read lots of industry reports telling me how unsuccessful Windows 8 has been, much worse even than the dreaded Vista. Even Microsoft with its huge marketing budget has not been able to buck the trend. Apple also reports lower sales of its PCs and the report suggests they may have been cannibalized by Apple’s own tablets (how ironic).

Is it all to do with the ongoing world financial crisis? Do we blame the politicians and bureaucrats of Ireland, Iceland, Spain, Portugal, Italy, Greece and now Cypress for this massive fall off in PC Shipments? Or, as I surmise, are we all more than a little bored with the IT industry, its hype and the too regular platform changes forced upon us? Are we all jaded by a decade of too rapid and unneeded change?

I like Windows 7, it works, it is stable and it allows me to run all the programs I need for my business. Why would I upgrade especially as I am going to have to retrain all my staff and also have to upgrade a lot of the software and hardware I use? What compelling reason is there to upgrade to Windows 8?

Similarly, my desktops and servers are now 3 to 4 years old but I bought high quality Dell OptiPlex PCs and Dell Xeon rack servers and they are all still more powerful than I need and still working fine. When something occasionally fails I just pay Dell to fix or replace it. It is a lot less disruptive and a lot less costly than replacing everything. What compelling reason is there for me to suffer the pain and disruption of replacing my PCs and servers?

Of course the world financial crisis has a lot to do with the tumbling PC sales figures because most organisations are still cutting costs to maintain or grow profits. However, I also detect a sea change in attitudes among my peer groups and customers. We have had enough of constant change for change’s sake. Most of the people I deal with are now sticking by the old maxim of “If it ain’t broke, don’t fix it.”

It looks like a lot of us have all lost interest in technology, we have even become bored and blasé about technology. So it is 10% lighter and 15% faster, “who cares?’ So it is prettier and has even more features I won’t ever use, “who cares?” There is another iPhone that is slightly bigger and slightly thinner than the last one, “who cares?” There is yet another update to Linux or Android, “who cares?”

I own and run a computer software company called Knowledgeone Corporation that builds and markets a range of enterprise content management software applications under the banner of RecFind 6. Because of this I am vitally interested in what is happening both with the ongoing world financial crisis and PC shipments because both affect my business.

Just like my customers, I am fed up with the industry trying to force feed me with new products that I don’t need and frankly, am just not interested in. I am the same as my customers, they just want my products to work day in and day out, 24/7, and do the job they were purchased for. They will buy maintenance because that protects their investment in my products but right now, most aren’t really ready to face or fund a massive change in their operation unless there is a damn good reason with a sound business justification.

I believe one of the main reasons PC sales are down, in addition to the world financial crisis, is because right now we just aren’t interested in new technology for technologies sake. We are more interested in running our businesses in the most cost effective manner and maintaining profitability. We are also tired of the IT industry trying to hard sell another ‘new thing’ every 3 years or so.

I don’t need new PCs, I don’t need new servers, I don’t need the next iPhone or update to Android. I think the world as a whole is now clearly differentiating between need and want and if need rather than want is driving the system then trying to woo us with faster, thinner, prettier technology just isn’t going to work. Frankly, I think we are bored with technology and all have more important things to think about like how to remain profitable and protect our companies and the jobs of our staff.

Maybe we are all waiting for the It industry to come up with something really, really interesting and really, really useful that will actually help us strengthen our bottom line? Now that would be something new.

Month List