Index: src/java/org/apache/nutch/fetcher/FetcherOutputFormat.java
===================================================================
--- src/java/org/apache/nutch/fetcher/FetcherOutputFormat.java	(revision 467730)
+++ src/java/org/apache/nutch/fetcher/FetcherOutputFormat.java	(working copy)
@@ -26,6 +26,7 @@
 import org.apache.hadoop.io.WritableComparable;
 import org.apache.hadoop.io.Writable;
 import org.apache.hadoop.io.Text;
+import org.apache.hadoop.io.SequenceFile.CompressionType;
 
 import org.apache.hadoop.mapred.OutputFormat;
 import org.apache.hadoop.mapred.RecordWriter;
@@ -55,7 +56,9 @@
       new Path(new Path(job.getOutputPath(), Content.DIR_NAME), name);
 
     final MapFile.Writer fetchOut =
-      new MapFile.Writer(fs, fetch.toString(), Text.class, CrawlDatum.class);
+      new MapFile.Writer(job, fs, fetch.toString(),
+                         Text.class, CrawlDatum.class,
+                         CompressionType.NONE, progress);
     
     return new RecordWriter() {
         private MapFile.Writer contentOut;
@@ -63,12 +66,14 @@
 
         {
           if (Fetcher.isStoringContent(job)) {
-            contentOut = new MapFile.Writer(fs, content.toString(),
-                                            Text.class, Content.class);
+            contentOut = new MapFile.Writer(job, fs, content.toString(),
+                                            Text.class, Content.class,
+                                            CompressionType.NONE, progress);
           }
 
           if (Fetcher.isParsing(job)) {
-            parseOut = new ParseOutputFormat().getRecordWriter(fs, job, name, null);
+            parseOut = new ParseOutputFormat().getRecordWriter
+              (fs, job, name, progress);
           }
         }
 
Index: src/java/org/apache/nutch/crawl/LinkDb.java
===================================================================
--- src/java/org/apache/nutch/crawl/LinkDb.java	(revision 467730)
+++ src/java/org/apache/nutch/crawl/LinkDb.java	(working copy)
@@ -315,10 +315,11 @@
     Path old = new Path(linkDb, "old");
     Path current = new Path(linkDb, CURRENT_NAME);
     fs.delete(old);
-    fs.rename(current, old);
+    if (fs.exists(current)) {
+      fs.rename(current, old);
+    }
     fs.mkdirs(linkDb);
     fs.rename(newLinkDb, current);
-    fs.delete(old);
   }
 
   public static void main(String[] args) throws Exception {
Index: src/java/org/apache/nutch/crawl/CrawlDb.java
===================================================================
--- src/java/org/apache/nutch/crawl/CrawlDb.java	(revision 467730)
+++ src/java/org/apache/nutch/crawl/CrawlDb.java	(working copy)
@@ -115,10 +115,11 @@
     Path old = new Path(crawlDb, "old");
     Path current = new Path(crawlDb, CrawlDatum.DB_DIR_NAME);
     fs.delete(old);
-    fs.rename(current, old);
+    if (fs.exists(current)) {
+      fs.rename(current, old);
+    }
     fs.mkdirs(crawlDb);
     fs.rename(newCrawlDb, current);
-    fs.delete(old);
   }
 
   public static void main(String[] args) throws Exception {
Index: src/java/org/apache/nutch/parse/ParseOutputFormat.java
===================================================================
--- src/java/org/apache/nutch/parse/ParseOutputFormat.java	(revision 467730)
+++ src/java/org/apache/nutch/parse/ParseOutputFormat.java	(working copy)
@@ -21,6 +21,8 @@
 import org.apache.commons.logging.LogFactory;
 
 import org.apache.hadoop.io.*;
+import org.apache.hadoop.util.Progressable;
+import org.apache.hadoop.io.SequenceFile.CompressionType;
 import org.apache.nutch.crawl.CrawlDatum;
 import org.apache.nutch.fetcher.Fetcher;
 import org.apache.hadoop.fs.*;
@@ -66,13 +68,16 @@
       new Path(new Path(job.getOutputPath(), CrawlDatum.PARSE_DIR_NAME), name);
     
     final MapFile.Writer textOut =
-      new MapFile.Writer(fs, text.toString(), Text.class, ParseText.class);
+      new MapFile.Writer(job, fs, text.toString(),Text.class,ParseText.class,
+                         CompressionType.NONE, progress);
     
     final MapFile.Writer dataOut =
-      new MapFile.Writer(fs, data.toString(), Text.class,ParseData.class,true);
+      new MapFile.Writer(job, fs, data.toString(),Text.class,ParseData.class,
+                         CompressionType.RECORD, progress);
     
     final SequenceFile.Writer crawlOut =
-      new SequenceFile.Writer(fs, crawl, Text.class, CrawlDatum.class);
+      SequenceFile.createWriter(fs, job, crawl, Text.class, CrawlDatum.class,
+                                CompressionType.NONE, progress);
     
     return new RecordWriter() {
 
