mirror of
				https://github.com/Alfresco/alfresco-community-repo.git
				synced 2025-10-29 15:21:53 +00:00 
			
		
		
		
	git-svn-id: https://svn.alfresco.com/repos/alfresco-enterprise/alfresco/HEAD/root@20925 c4b6b30b-aa2e-2d43-bbcb-ca4b014f7261
		
			
				
	
	
		
			256 lines
		
	
	
		
			9.6 KiB
		
	
	
	
		
			Java
		
	
	
	
	
	
			
		
		
	
	
			256 lines
		
	
	
		
			9.6 KiB
		
	
	
	
		
			Java
		
	
	
	
	
	
| /*
 | |
|  * Copyright (C) 2005-2010 Alfresco Software Limited.
 | |
|  *
 | |
|  * This file is part of Alfresco
 | |
|  *
 | |
|  * Alfresco is free software: you can redistribute it and/or modify
 | |
|  * it under the terms of the GNU Lesser General Public License as published by
 | |
|  * the Free Software Foundation, either version 3 of the License, or
 | |
|  * (at your option) any later version.
 | |
|  *
 | |
|  * Alfresco is distributed in the hope that it will be useful,
 | |
|  * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
|  * GNU Lesser General Public License for more details.
 | |
|  *
 | |
|  * You should have received a copy of the GNU Lesser General Public License
 | |
|  * along with Alfresco. If not, see <http://www.gnu.org/licenses/>.
 | |
|  */
 | |
| package org.alfresco.repo.content.metadata;
 | |
| 
 | |
| import java.io.BufferedInputStream;
 | |
| import java.io.File;
 | |
| import java.io.FileInputStream;
 | |
| import java.io.Serializable;
 | |
| import java.net.URL;
 | |
| import java.util.ArrayList;
 | |
| import java.util.HashMap;
 | |
| import java.util.HashSet;
 | |
| import java.util.Map;
 | |
| import java.util.Set;
 | |
| 
 | |
| import org.alfresco.model.ContentModel;
 | |
| import org.alfresco.repo.content.filestore.FileContentReader;
 | |
| import org.alfresco.repo.content.transform.AbstractContentTransformerTest;
 | |
| import org.alfresco.service.cmr.repository.ContentReader;
 | |
| import org.alfresco.service.cmr.repository.datatype.DefaultTypeConverter;
 | |
| import org.alfresco.service.namespace.QName;
 | |
| import org.apache.tika.metadata.Metadata;
 | |
| import org.apache.tika.mime.MediaType;
 | |
| import org.apache.tika.parser.AutoDetectParser;
 | |
| import org.apache.tika.parser.ParseContext;
 | |
| import org.apache.tika.parser.Parser;
 | |
| import org.apache.tika.parser.microsoft.OfficeParser;
 | |
| import org.apache.tika.parser.microsoft.ooxml.OOXMLParser;
 | |
| import org.apache.tika.parser.mp3.Mp3Parser;
 | |
| import org.apache.tika.parser.odf.OpenDocumentParser;
 | |
| 
 | |
| 
 | |
| /**
 | |
|  * @see TikaAutoMetadataExtracter
 | |
|  * 
 | |
|  * @author Nick Burch
 | |
|  */
 | |
| public class TikaAutoMetadataExtracterTest extends AbstractMetadataExtracterTest
 | |
| {
 | |
|     private TikaAutoMetadataExtracter extracter;
 | |
|     private static final QName TIKA_MIMETYPE_TEST_PROPERTY =
 | |
|        QName.createQName("TikaMimeTypeTestProp");
 | |
| 
 | |
|     @Override
 | |
|     public void setUp() throws Exception
 | |
|     {
 | |
|         super.setUp();
 | |
|         extracter = new TikaAutoMetadataExtracter();
 | |
|         extracter.setDictionaryService(dictionaryService);
 | |
|         extracter.register();
 | |
|         
 | |
|         // Attach some extra mappings, using the Tika
 | |
|         //  metadata keys namespace
 | |
|         // These will be tested later
 | |
|         HashMap<String, Set<QName>> newMap = new HashMap<String, Set<QName>>(
 | |
|               extracter.getMapping()
 | |
|         );
 | |
|         
 | |
|         Set<QName> tlaSet = new HashSet<QName>();
 | |
|         tlaSet.add(TIKA_MIMETYPE_TEST_PROPERTY);
 | |
|         newMap.put( Metadata.CONTENT_TYPE, tlaSet );
 | |
|         
 | |
|         extracter.setMapping(newMap);
 | |
|     }
 | |
| 
 | |
|     /**
 | |
|      * @return Returns the same transformer regardless - it is allowed
 | |
|      */
 | |
|     protected MetadataExtracter getExtracter()
 | |
|     {
 | |
|         return extracter;
 | |
|     }
 | |
| 
 | |
|     public void testSupports() throws Exception
 | |
|     {
 | |
|         ArrayList<String> mimeTypes = new ArrayList<String>();
 | |
|         for (Parser p : new Parser[] {
 | |
|                  new OfficeParser(), new OpenDocumentParser(),
 | |
|                  new Mp3Parser(), new OOXMLParser()
 | |
|         }) {
 | |
|            Set<MediaType> mts = p.getSupportedTypes(new ParseContext());
 | |
|            for (MediaType mt : mts) {
 | |
|               mimeTypes.add(mt.toString());
 | |
|            }
 | |
|         }
 | |
|         
 | |
|         for (String mimetype : mimeTypes)
 | |
|         {
 | |
|             boolean supports = extracter.isSupported(mimetype);
 | |
|             assertTrue("Mimetype should be supported: " + mimetype, supports);
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     /**
 | |
|      * Test several different files
 | |
|      * Note - doesn't use extractFromMimetype
 | |
|      */
 | |
|     public void testSupportedMimetypes() throws Exception
 | |
|     {
 | |
|         String[] testFiles = new String[] {
 | |
|               ".doc", ".docx", ".xls", ".xlsx",
 | |
|               ".ppt", ".pptx", 
 | |
|               //".vsd", // Not auto-detected properly yet
 | |
|               //"2010.dwg", // Not auto-detected properly yet
 | |
|               ".pdf",
 | |
|               ".odt"
 | |
|         };
 | |
|            
 | |
|         for (String fileBase : testFiles)
 | |
|         {
 | |
|            String filename = "quick" + fileBase;
 | |
|            URL url = AbstractContentTransformerTest.class.getClassLoader().getResource("quick/" + filename);
 | |
|            File file = new File(url.getFile());
 | |
|            
 | |
|            // Cheat and ask Tika for the mime type!
 | |
|            AutoDetectParser ap = new AutoDetectParser();
 | |
|            Metadata metadata = new Metadata();
 | |
|            metadata.set(Metadata.RESOURCE_NAME_KEY, filename);
 | |
|            MediaType mt = ap.getDetector().detect(
 | |
|                  new BufferedInputStream(new FileInputStream(file)), metadata);
 | |
|            String mimetype = mt.toString();
 | |
| 
 | |
|            // Have it processed
 | |
|            Map<QName, Serializable> properties = extractFromFile(file, mimetype);
 | |
|            
 | |
|            // check we got something
 | |
|            assertFalse("extractFromMimetype should return at least some properties, " +
 | |
|            		"none found for " + mimetype + " - " + filename,
 | |
|               properties.isEmpty());
 | |
|            
 | |
|            // check common metadata
 | |
|            testCommonMetadata(mimetype, properties);
 | |
|            // check file-type specific metadata
 | |
|            testFileSpecificMetadata(mimetype, properties);
 | |
|         }
 | |
|     }
 | |
|     
 | |
|     @Override
 | |
|     protected boolean skipAuthorCheck(String mimetype) { return true; }
 | |
| 
 | |
|    /**
 | |
|     * We also provide the creation date - check that
 | |
|     */
 | |
|    protected void testFileSpecificMetadata(String mimetype,
 | |
|          Map<QName, Serializable> properties) {
 | |
|       
 | |
|       // Check for extra fields
 | |
|       // Author isn't there for the OpenDocument ones
 | |
|       if(mimetype.indexOf(".oasis.") == -1) {
 | |
|          assertEquals(
 | |
|                "Property " + ContentModel.PROP_AUTHOR + " not found for mimetype " + mimetype,
 | |
|                "Nevin Nollop",
 | |
|                DefaultTypeConverter.INSTANCE.convert(String.class, properties.get(ContentModel.PROP_AUTHOR)));
 | |
|       }
 | |
|       
 | |
|       // Ensure that we can also get things which are standard
 | |
|       //  Tika metadata properties, if we so choose to
 | |
|       assertTrue( 
 | |
|             "Test Property " + TIKA_MIMETYPE_TEST_PROPERTY + " not found for mimetype " + mimetype,
 | |
|             properties.containsKey(TIKA_MIMETYPE_TEST_PROPERTY)
 | |
|       );
 | |
|       // TODO - uncomment this when TIKA-391 is properly fixed
 | |
| //      assertEquals(
 | |
| //            "Test Property " + TIKA_MIMETYPE_TEST_PROPERTY + " incorrect for mimetype " + mimetype,
 | |
| //            mimetype,
 | |
| //            DefaultTypeConverter.INSTANCE.convert(String.class, properties.get(TIKA_MIMETYPE_TEST_PROPERTY)));
 | |
|    }
 | |
| 
 | |
|    /**
 | |
|     * We don't have explicit extractors for most image and video formats.
 | |
|     * Instead, these will be handled by the Auto Tika Parser, and
 | |
|     *  this test ensures that they are
 | |
|     */
 | |
|    public void testImageVideo() throws Throwable {
 | |
|       Map<String, Serializable> p;
 | |
|       
 | |
|       // Image
 | |
|       p = openAndCheck(".jpg", "image/jpeg");
 | |
|       assertEquals("409 pixels", p.get("Image Width"));
 | |
|       assertEquals("92 pixels", p.get("Image Height"));
 | |
|       assertEquals("8 bits", p.get("Data Precision"));
 | |
|       
 | |
|       p = openAndCheck(".gif", "image/gif");
 | |
|       assertEquals("409", p.get("width"));
 | |
|       assertEquals("92", p.get("height"));
 | |
|       
 | |
|       p = openAndCheck(".png", "image/png");
 | |
|       assertEquals("409", p.get("width"));
 | |
|       assertEquals("92", p.get("height"));
 | |
|       assertEquals("8 8 8", p.get("Data BitsPerSample"));
 | |
|       assertEquals("none", p.get("Transparency Alpha"));
 | |
|       
 | |
|       p = openAndCheck(".bmp", "image/bmp");
 | |
|       assertEquals("409", p.get("width"));
 | |
|       assertEquals("92", p.get("height"));
 | |
|       assertEquals("8 8 8", p.get("Data BitsPerSample"));
 | |
|       
 | |
|       
 | |
|       // Geo tagged image
 | |
|       p = openAndCheck("GEO.jpg", "image/jpeg");
 | |
|       assertEquals("100 pixels", p.get("Image Width"));
 | |
|       assertEquals("68 pixels", p.get("Image Height"));
 | |
|       assertEquals("8 bits", p.get("Data Precision"));
 | |
|       assertEquals(QUICK_TITLE, p.get("Comments"));
 | |
|       assertEquals("12.54321", p.get("geo:lat"));
 | |
|       assertEquals("-54.1234", p.get("geo:long"));
 | |
|       
 | |
|       // Map and check
 | |
|       Map<QName, Serializable> properties = new HashMap<QName, Serializable>();
 | |
|       ContentReader reader = new FileContentReader(open("GEO.jpg"));
 | |
|       reader.setMimetype("image/jpeg");
 | |
|       extracter.extract(reader, properties);
 | |
|       assertEquals(12.54321, properties.get(QName.createQName("http://www.alfresco.org/model/content/1.0","latitude")));
 | |
|       assertEquals(-54.1234, properties.get(QName.createQName("http://www.alfresco.org/model/content/1.0","longitude")));
 | |
|    }
 | |
|    private File open(String fileBase) throws Throwable {
 | |
|       String filename = "quick" + fileBase;
 | |
|       URL url = AbstractContentTransformerTest.class.getClassLoader().getResource("quick/" + filename);
 | |
|       File file = new File(url.getFile());
 | |
|       assertTrue(file.exists());
 | |
|       return file;
 | |
|    }
 | |
|    private Map<String, Serializable> openAndCheck(String fileBase, String expMimeType) throws Throwable {
 | |
|       // Cheat and ask Tika for the mime type!
 | |
|       File file = open(fileBase);
 | |
|       AutoDetectParser ap = new AutoDetectParser();
 | |
|       Metadata metadata = new Metadata();
 | |
|       metadata.set(Metadata.RESOURCE_NAME_KEY, "quick"+fileBase);
 | |
|       MediaType mt = ap.getDetector().detect(
 | |
|             new BufferedInputStream(new FileInputStream(file)), metadata);
 | |
|       String mimetype = mt.toString();
 | |
| 
 | |
|       assertEquals("Wrong mimetype for " + fileBase, mimetype, expMimeType);
 | |
|       
 | |
|       ContentReader sourceReader = new FileContentReader(file);
 | |
|       sourceReader.setMimetype(mimetype);
 | |
|       return extracter.extractRaw(sourceReader);
 | |
|    }
 | |
| }
 |