blob: d66660fb5f64b78255130da9ab638bfe2269644f [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*
*/
package org.apache.directory.server.tools;
import java.io.File;
import java.io.FileWriter;
import java.io.PrintWriter;
import java.util.ArrayList;
import java.util.HashSet;
import java.util.List;
import java.util.Set;
import javax.naming.NamingEnumeration;
import javax.naming.directory.Attribute;
import javax.naming.directory.Attributes;
import jdbm.helper.MRU;
import jdbm.recman.BaseRecordManager;
import jdbm.recman.CacheRecordManager;
import org.apache.commons.cli.CommandLine;
import org.apache.commons.cli.Option;
import org.apache.commons.cli.Options;
import org.apache.directory.server.core.DefaultDirectoryService;
import org.apache.directory.server.core.DirectoryService;
import org.apache.directory.server.core.entry.ServerEntry;
import org.apache.directory.server.core.partition.impl.btree.jdbm.JdbmIndex;
import org.apache.directory.server.core.partition.impl.btree.jdbm.JdbmMasterTable;
import org.apache.directory.server.xdbm.Tuple;
import org.apache.directory.shared.ldap.MultiException;
import org.apache.directory.shared.ldap.cursor.Cursor;
import org.apache.directory.shared.ldap.exception.LdapConfigurationException;
import org.apache.directory.shared.ldap.ldif.LdifUtils;
import org.apache.directory.shared.ldap.schema.AttributeType;
import org.apache.directory.shared.ldap.schema.SchemaManager;
import org.apache.directory.shared.ldap.schema.UsageEnum;
import org.apache.directory.shared.ldap.schema.ldif.extractor.SchemaLdifExtractor;
import org.apache.directory.shared.ldap.schema.ldif.extractor.impl.DefaultSchemaLdifExtractor;
import org.apache.directory.shared.ldap.schema.loader.ldif.LdifSchemaLoader;
import org.apache.directory.shared.ldap.schema.manager.impl.DefaultSchemaManager;
import org.apache.directory.shared.ldap.util.Base64;
import org.apache.directory.shared.ldap.util.ExceptionUtils;
/**
* Simple tool used to dump the contents of a jdbm based partition.
*
* @author <a href="mailto:dev@directory.apache.org">Apache Directory Project</a>
* @version $Rev: 493916 $
*/
public class DumpCommand extends ToolCommand
{
private SchemaManager schemaManager;
private Set<String> exclusions = new HashSet<String>();
private boolean includeOperational = false;
public DumpCommand()
{
super( "dump" );
}
private SchemaManager loadSchemaManager() throws Exception
{
// --------------------------------------------------------------------
// Load the bootstrap schemas to start up the schema partition
// --------------------------------------------------------------------
// setup temporary loader and temp registry
String workingDirectory = System.getProperty( "workingDirectory" );
if ( workingDirectory == null )
{
String path = DumpCommand.class.getResource( "" ).getPath();
int targetPos = path.indexOf( "target" );
workingDirectory = path.substring( 0, targetPos + 6 );
}
File schemaRepository = new File( workingDirectory, "schema" );
SchemaLdifExtractor extractor = new DefaultSchemaLdifExtractor( new File( workingDirectory ) );
extractor.extractOrCopy();
LdifSchemaLoader loader = new LdifSchemaLoader( schemaRepository );
schemaManager = new DefaultSchemaManager( loader );
schemaManager.loadAllEnabled();
List<Throwable> errors = schemaManager.getErrors();
if ( errors.size() != 0 )
{
throw new Exception( "Schema load failed : " + ExceptionUtils.printErrors( errors ) );
}
schemaManager.loadWithDeps( "collective" );
errors = schemaManager.getErrors();
if ( !errors.isEmpty() )
{
MultiException e = new MultiException();
for ( Throwable t : errors )
{
e.addThrowable( t );
}
throw e;
}
// --------------------------------------------------------------------
// Initialize schema partition or bomb out if we cannot find it on disk
// --------------------------------------------------------------------
// If not present then we need to abort
File schemaDirectory = new File( getInstanceLayout().getPartitionsDir(), "schema" );
if ( !schemaDirectory.exists() )
{
throw new LdapConfigurationException( "The following schema directory from "
+ "the installation layout could not be found:\n\t" + schemaDirectory );
}
DirectoryService directoryService = new DefaultDirectoryService();
//schemaPartition.init( directoryService );
// --------------------------------------------------------------------
// Initialize schema subsystem and reset registries
// --------------------------------------------------------------------
// PartitionSchemaLoader schemaLoader = new PartitionSchemaLoader( schemaPartition, registries );
// schemaLoader.loadEnabled( globalRegistries );
// SerializableComparator.setRegistry( globalRegistries.getComparatorRegistry() );
return schemaManager;
}
public void execute( CommandLine cmdline ) throws Exception
{
getLayout().verifyInstallation();
schemaManager = loadSchemaManager();
includeOperational = cmdline.hasOption( 'o' );
String[] partitions = cmdline.getOptionValues( 'p' );
String outputFile = cmdline.getOptionValue( 'f' );
PrintWriter out = null;
String[] excludedAttributes = cmdline.getOptionValues( 'e' );
if ( excludedAttributes != null )
{
for ( String attributeType:excludedAttributes)
{
AttributeType type = schemaManager.lookupAttributeTypeRegistry( attributeType );
exclusions.add( type.getName() );
}
}
if ( outputFile == null )
{
out = new PrintWriter( System.out );
}
else
{
out = new PrintWriter( new FileWriter( outputFile ) );
}
for ( String partition:partitions )
{
File partitionDirectory = new File( getInstanceLayout().getPartitionsDir(), partition );
out.println( "\n\n" );
dump( partitionDirectory, out );
}
}
private void dump( File partitionDirectory, PrintWriter out ) throws Exception
{
if ( !partitionDirectory.exists() )
{
System.err.println( "Partition directory " + partitionDirectory + " does not exist!" );
System.exit( 1 );
}
out.println( "# ========================================================================" );
out.println( "# ApacheDS Tools Version: " + getVersion() );
out.println( "# Partition Directory: " + partitionDirectory );
out.println( "# ========================================================================\n\n" );
String path = partitionDirectory.getPath() + File.separator + "master";
BaseRecordManager base = new BaseRecordManager( path );
base.disableTransactions();
CacheRecordManager recMan = new CacheRecordManager( base, new MRU( 1000 ) );
JdbmMasterTable<ServerEntry> master = new JdbmMasterTable<ServerEntry>( recMan, schemaManager );
AttributeType attributeType = schemaManager.lookupAttributeTypeRegistry( "apacheUpdn" );
JdbmIndex idIndex = new JdbmIndex();
idIndex.setAttributeId( attributeType.getName() );
idIndex.setWkDirPath( partitionDirectory );
idIndex.setCacheSize( 1000 );
idIndex.setNumDupLimit( 1000 );
idIndex.init( schemaManager, attributeType, partitionDirectory );
out.println( "#---------------------" );
Cursor<Tuple<Long,ServerEntry>> list = master.cursor();
StringBuffer buf = new StringBuffer();
while ( list.next() )
{
Tuple<Long,ServerEntry> tuple = list.get();
Long id = tuple.getKey();
String dn = ( String ) idIndex.reverseLookup( id );
Attributes entry = ( Attributes ) tuple.getValue();
filterAttributes( dn, entry );
buf.append( "# Entry: " ).append( id ).append( "\n#---------------------\n\n" );
if ( !LdifUtils.isLDIFSafe( dn ) )
{
// If the DN isn't LdifSafe, it needs to be Base64 encoded.
buf.append( "dn:: " ).append( new String( Base64.encode( dn.getBytes() ) ) );
}
else
{
buf.append( "dn: " ).append( dn );
}
buf.append( "\n" ).append( LdifUtils.convertToLdif( entry ) );
if ( list.next() )
{
buf.append( "\n\n#---------------------\n" );
}
out.print( buf.toString() );
out.flush();
buf.setLength( 0 );
}
}
private void filterAttributes( String dn, Attributes entry ) throws Exception
{
List<String> toRemove = new ArrayList<String>();
NamingEnumeration<? extends Attribute> attrs = entry.getAll();
while ( attrs.hasMore() )
{
Attribute attr = attrs.next();
if ( !schemaManager.getAttributeTypeRegistry().contains( attr.getID() ) )
{
if ( !isQuietEnabled() )
{
System.out
.println( "# Cannot properly filter unrecognized attribute " + attr.getID() + " in " + dn );
}
continue;
}
AttributeType type = schemaManager.lookupAttributeTypeRegistry( attr.getID() );
boolean isOperational = type.getUsage() != UsageEnum.USER_APPLICATIONS;
if ( exclusions.contains( attr.getID() ) || ( isOperational && ( !includeOperational ) ) )
{
toRemove.add( attr.getID() );
}
}
for ( String id:toRemove )
{
entry.remove( id );
if ( isDebugEnabled() )
{
System.out.println( "# Excluding attribute " + id + " in " + dn );
}
}
}
public Options getOptions()
{
Options opts = new Options();
Option op = new Option( "f", "file", true, "file to output the dump to" );
op.setRequired( false );
opts.addOption( op );
op = new Option( "p", "partitions", true, "the partitions to dump" );
op.setRequired( true );
op.setValueSeparator( File.pathSeparatorChar );
opts.addOption( op );
op = new Option( "e", "excluded-attributes", true, "the attributes to exclude" );
op.setRequired( false );
op.setValueSeparator( File.pathSeparatorChar );
opts.addOption( op );
op = new Option( "o", "include-operational", false, "include operational attributes: defaults to false" );
op.setRequired( false );
opts.addOption( op );
op = new Option( "i", "install-path", true, "path to apacheds installation directory" );
op.setRequired( true );
opts.addOption( op );
return opts;
}
}