Development/Tutorials/Writing file analyzers

    From KDE TechBase
    Revision as of 09:19, 14 March 2007 by Dfaure (talk | contribs) (ok |= doesn't perform shortcut evaluation; ok = ok || does, and is therefore faster.)

    Writing KDE4 file analyzers

    File analyzers extract data from files to display in the file dialogs and file managers. The data gathered this way is also used to search for files. KDE4 allows the use of multiple analyzers per file type. Analyzers can extract text which is used for indexing, but they can also retrieve other data such as song title, album title, recipient, md5 sum, the mimetype of a file, and much more.

    This tutorial describes how you can write new analyzers.

    Primer

    What are file analyzers?

    A file analyzer is a class that extracts metadata from a file or data stream. You can have file analyzers that are specific for certain file types such as an analyzer that extracts the information from an ogg vorbis file. There are also more general file analyzers that calculate for example the md5 or sha1 of a file.

    File analyzers in KDE4

    KDE4 uses stream based file analyzers for retrieving text and metadata from files. This has a number of advantages over file based methods. Stream based access

    • is faster for 90% of the file types,
    • allows easy analysis of embedded files such as email attachments or entries from zip files, rpms and many other container file formats.

    Writing stream-based analyzers requires a different approach than the usual file-based methods and in the tutorial we will explain how to go about it.


    Look for existing code

    If you want to see some code examples, take a look at the already implemented file analyzers at /kdesupport/strigi/src/streamindexer/

    Choosing the type of analyzer

    There are two types of analyzers: StreamThroughAnalyzer and StreamEndAnalyzer. The latter is more powerful and a bit easier to program, but has a limition: only one StreamEndAnalyzer can analyze a particular resource, while you can use as many StreamThroughAnalyzers as you like. Most analyzers can be written as StreamThroughAnalyzers. The most important exception is for analyzers that extract embedded resources from a stream. Examples of this are the ZipEndAnalyzer, the MailEndAnalyzer and the RpmEndAnalyzer.

    In this tutorial we focus on a simple example file type: BMP images. The information we will get from this file is located at the start of the file. It turns out that in this case, it is just as easy to implement the analyzer as a StreamEndAnalyzer as a StreamThroughAnalyzer. We will implement it as a StreamEndAnalyzer and point out how to do the same as a StreamThroughAnalyzer.

    StreamEndAnalyzer

    Three functions need to be implemented in a StreamEndAnalyzer. The function of getName() is obvious so we will discuss only two: checkHeader() and analyze().

    class BmpEndAnalyzer : public jstreams::StreamEndAnalyzer { private:

       const BmpEndAnalyzerFactory* factory;
    

    public:

       BmpEndAnalyzer(const BmpEndAnalyzerFactory* f) :factory(f) {}
       bool checkHeader(const char* header, int32_t headersize) const;
       char analyze(jstreams::Indexable& idx, jstreams::InputStream* in);
       const char* getName() const { return "BmpEndAnalyzer"; }
    

    };

    Since only one StreamEndAnalyzer can be used per resource, it is important to quickly select the right one. For this, we dont rely on the mimetype, but on the actual contents of the resource. As a first sifting, the initial bytes of a file are checked. This is usually just as fast as comparing a mimetype identifier, but has the advantage of being more direct and thus often more accurate.

    It is by no means necessary for checkHeader to be 100% correct. The most important thing is that it should allow one to quickly determine if an analyzer can not handle a resource. Should it by accident return true and thus indicate that a resource can be handled, then it can always handle this in the analyze() method.

    bool BmpEndAnalyzer::checkHeader(const char* header, int32_t headersize) const {

       bool ok = false;
       if (headersize > 2) {
           ok = ok || (strncmp(header, "BM", 2) == 0);
           ok = ok || (strncmp(header, "BA", 2) == 0);
           ok = ok || (strncmp(header, "CI", 2) == 0);
           ok = ok || (strncmp(header, "CP", 2) == 0);
           ok = ok || (strncmp(header, "IC", 2) == 0);
           ok = ok || (strncmp(header, "PT", 2) == 0);
       }
       return ok;
    

    }

    A BMP file can start with six different initial bytes. If the header matches any of them, we return true.

    If a resource passes this test, the analyze() function will be called. The real work occurs in this function. In this tutorial, we will not do a complete analysis, but we only look at the way the pixels are stored in the BMP file. This information is stored in the bytes 30-33 that encode a number. char BmpEndAnalyzer::analyze(Indexable& idx, InputStream* in) {

       // read compression type (bytes #30-33)
       const char* h;
       int32_t n = in->read(h, 34, 34); // read exactly 34 bytes
       in->reset(0);   // rewind to the start of the stream
       if (n < 34) return Error;
    
       uint32_t bmpi_compression = (unsigned char)h[33] + ((unsigned char)h[32]<<8)
            + ((unsigned char)h[31]<<16) + ((unsigned char)h[30]<<24);
    
       switch (bmpi_compression) {
       case 0 :
           idx.setField(factory->compressionField, "None");
           break;
       case 1 :
           idx.setField(factory->compressionField, "RLE 8bit/pixel");
           break;
       case 2 :
           idx.setField(factory->compressionField, "RLE 4bit/pixel");
           break;
       case 3 :
           idx.setField(factory->compressionField, "Bitfields");
           break;
       default :
           idx.setField(factory->compressionField, "Unknown");
       }
       return Ok;
    

    } First we read exactly 34 bytes from the stream. We do not need to allocate a buffer; this is handled by the stream. The stream returns a pointer to its internal buffer, which avoids data copying and buffer allocation. If the resource has less than 34 bytes or if an error occurred during reading, we return with an error code.

    The bytes 30-33 contain the information we need and we write this into the Indexable object. This object collects all metadata and passes it to the code that initiated the analysis. The data may get written into an index or get passed to for example a file dialog. To indicate the type of metadata, we pass a pointer to a registered field.

    StreamThroughAnalyzer

    We will now look at how to implement a StreamThroughAnalyzer. For this, we need to implement three functions. class BmpThroughAnalyzer : public jstreams::StreamThroughAnalyzer { private:

       jstreams::Indexable* indexable;
       const BmpThroughAnalyzerFactory* factory;
    

    public:

       BmpThroughAnalyzer(const BmpThroughAnalyzerFactory* f) :factory(f) {}
       ~BmpThroughAnalyzer() {}
       void setIndexable(jstreams::Indexable* i) { indexable = i; }
       jstreams::InputStream *connectInputStream(jstreams::InputStream *in);
       bool isReadyWithStream() { return true; }
    

    };

    For simple file formats where all information is in the initial part of the file, all of the work gets done in connectInputStream(). For more complicated cases look at other examples, such as DigestThroughAnalyzer. In connectInputStream(), we perform the same analysis as we did in the StreamEndAnalyzer: jstreams::InputStream BmpThroughAnalyzer::connectInputStream(InputStream* in) { // read compression type (bytes #30-33)

       const char* h;
       int32_t n = in->read(h, 34, 34); // read exactly 34 bytes
       in->reset(0);   // rewind to the start of the stream
       if (n < 34) return in;
    
       uint32_t bmpi_compression = (unsigned char)h[33] + ((unsigned char)h[32]<<8)
            + ((unsigned char)h[31]<<16) + ((unsigned char)h[30]<<24);
    
       switch (bmpi_compression) {
       case 0 :
           indexable->setField(factory->compressionField, "None");
           break;
       case 1 :
           indexable->setField(factory->compressionField, "RLE 8bit/pixel");
           break;
       case 2 :
           indexable->setField(factory->compressionField, "RLE 4bit/pixel");
           break;
       case 3 :
           indexable->setField(factory->compressionField, "Bitfields");
           break;
       default :
           indexable->setField(factory->compressionField, "Unknown");
       }
       return in;
    

    } The difference with StreamEndAnalyzer is that we must make sure the only return a stream that has the current position at the start of the stream. We can ensure this by calling reset(0) after each call to read(). This will not affect the pointer to the character data that was set in the call to read().

    In addition, we do not return any status message, but an InputStream. In this case this is the same stream. For more complicated cases, we can subclass the InputStream and analyze the data that passes through it. This is where the StreamThroughAnalyzer gets its name.

    Since we have finished with the stream after the call to connectInputStream, we implement isReadyWithStream() to return true. This function is used by StreamIndexer to stop reading as soon as possible to speed up the analysis.

    Registering the analyzer

    KDE4 keeps a register of the capabilities of each analyzer. This allows it to speed up determining which analyzers to use. In addition, when it knows what data type an analyzer provides under what name, it can use this information to optimize the storage of the data or search queries. For this, each loadable analyzer must define two factories. An AnalyzerFactoryFactory and either a StreamThroughAnalyzerFactory or a StreamEndAnalyzerFactory.

    AnalyzerFactoryFactory

    The AnalyzerFactoryFactory is used only in the loadable plugins. It is not needed for analyzers that are part of the Strigi core. To initialize the StreamThroughAnalyzerFactories and StreamEndAnalyzerFactories in plugins, we need to implement an AnalyzerFactoryFactory. We do so by implementing one or two functions: getStreamThroughAnalyzerFactories() and getStreamEndAnalyzerFactories(). This function returns instances of all the factories available in a plugin. Here is for example the AnalyzerFactoryFactory for the KDE trash file analyzer (we cannot use the BMP analyzer here, because it is in Strigi core): class Factory : public AnalyzerFactoryFactory { public:

       list<StreamAnalyzerFactory*>
       getStreamThroughAnalyzerFactories() const {
           list<StreamThroughAnalyzerFactory*> af;
           af.push_back(new TrashThroughAnalyzerFactory());
           return af;
       }
    

    };

    // macro that initializes the Factory when the plugin is loaded STRIGI_ANALYZER_FACTORY(Factory)

    StreamEndAnalyzerFactory

    StreamEndAnalyzerFactories and StreamThroughAnalyzerFactories are very similar. They provide information about analyzers and the create instances of the analyzers. Each analyzer must hava a factory. StreamThroughAnalyzers have StreamThroughAnalyzerFactories and StreamEndAnalyzers have StreamEndAnalyzerFactories. Here, we look at the factory for the BmpEndAnalyzer.

    The class BmpEndAnalyzerFactory looks like this:

    class BmpEndAnalyzerFactory : public jstreams::StreamEndAnalyzerFactory { friend class BmpEndAnalyzer; private:

       static const cnstr typeFieldName;
       static const cnstr compressionFieldName;
       const jstreams::RegisteredField* typeField;
       const jstreams::RegisteredField* compressionField;
       const char* getName() const {
           return "BmpEndAnalyzer";
       }
       jstreams::StreamEndAnalyzer* newInstance() const {
           return new BmpEndAnalyzer(this);
       }
       void registerFields(jstreams::FieldRegister&);
    

    };

    All members are private, which is ok, because the important functions are virtual and thus accessible anyway. The functions getName() and newInstance() are selfexplanatory. The other important function is registerFields(jstreams::FieldRegister&). To speed up the extraction of fields, we dont use strings to identify fiels, but we use pointers to registered fields. These RegisteredField instances are stored in a global register. When you extract a piece of metadata, you pass the pointer to the registered field to identify the metadata.

    But first we need to register the fields:

    void BmpEndAnalyzerFactory::registerFields(FieldRegister& reg) {

       typeField = reg.registerField(typeFieldName, FieldRegister::stringType,
           1, 0);
       compressionField = reg.registerField(compressionFieldName,
           FieldRegister::stringType, 1, 0);
    

    }

    We pass the key of the field to the register, along with it's type, the maximum number of times the field occurs per resource and the parent of the field. The parent of a song tile, for example, could be a more general title field. The datatype of this field should be the same or a subset of the fieldtype of the parent.

    The pointers to the registered fields are used during the analysis to identify the type of data we have analyzed:

      idx.setField(factory->typeField, "OS/2 Color Icon");
    

    Building the analyzer

    Building your analyzer is easy. There are three things you must take into account:

    • Link the analyzer as a module,
    • Let the name of the analyzer start with 'strigita_' for a StreamThroughAnalyzer and 'strigiea_' for a StreamEndAnalyzer,
    • Install the plugin in the lib/strigi directory.

    add_library(trash MODULE trashthroughanalyzer.cpp trashimpl.cpp) target_link_libraries(trash ${STREAMINDEXER_LIBRARY} ${KDE4_KIO_LIBS} ${KDE4_SOLID_LIBS}) set_target_properties(trash PROPERTIES

       PREFIX strigita_)
    

    install(TARGETS trash LIBRARY DESTINATION ${LIB_DESTINATION}/strigi)

    Testing your code

    Strigi comes with a simple command line tool to check if your plugins work. This tool is called xmlindexer. It extracts data from files and outputs it as simple xml. To use it call it like this:

    xmlindexer [FILE]
    

    or

    xmlindexer [DIR]
    

    This is very fast and I recommend using it with valgrind. This hardly slows down your workflow but helps to keep memory management in good shape:

    valgrind xmlindexer [DIR]