You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
4778 lines
158 KiB
4778 lines
158 KiB
/* This file is part of the KDE libraries
|
|
Copyright (C) 2000 Stephan Kulow <coolo@kde.org>
|
|
David Faure <faure@kde.org>
|
|
Waldo Bastian <bastian@kde.org>
|
|
|
|
This library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Library General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2 of the License, or (at your option) any later version.
|
|
|
|
This library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Library General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Library General Public License
|
|
along with this library; see the file COPYING.LIB. If not, write to
|
|
the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor,
|
|
Boston, MA 02110-1301, USA.
|
|
*/
|
|
|
|
#include "kio/job.h"
|
|
|
|
#include <config.h>
|
|
|
|
#include <sys/types.h>
|
|
#include <sys/wait.h>
|
|
#include <sys/stat.h>
|
|
|
|
#include <assert.h>
|
|
|
|
#include <signal.h>
|
|
#include <stdlib.h>
|
|
#include <stdio.h>
|
|
#include <time.h>
|
|
#include <unistd.h>
|
|
extern "C" {
|
|
#include <pwd.h>
|
|
#include <grp.h>
|
|
}
|
|
#include <tqtimer.h>
|
|
#include <tqfile.h>
|
|
|
|
#include <kapplication.h>
|
|
#include <kglobal.h>
|
|
#include <klocale.h>
|
|
#include <ksimpleconfig.h>
|
|
#include <kdebug.h>
|
|
#include <kdialog.h>
|
|
#include <kmessagebox.h>
|
|
#include <kdatastream.h>
|
|
#include <kmainwindow.h>
|
|
#include <kde_file.h>
|
|
|
|
#include <errno.h>
|
|
|
|
#include "kmimetype.h"
|
|
#include "slave.h"
|
|
#include "scheduler.h"
|
|
#include "kdirwatch.h"
|
|
#include "kmimemagic.h"
|
|
#include "kprotocolinfo.h"
|
|
#include "kprotocolmanager.h"
|
|
|
|
#include "kio/observer.h"
|
|
|
|
#include "kssl/ksslcsessioncache.h"
|
|
|
|
#include <kdirnotify_stub.h>
|
|
#include <ktempfile.h>
|
|
#include <dcopclient.h>
|
|
|
|
#ifdef Q_OS_UNIX
|
|
#include <utime.h>
|
|
#endif
|
|
#if defined Q_WS_X11
|
|
#include <netwm.h>
|
|
#include <fixx11h.h>
|
|
#endif
|
|
|
|
using namespace KIO;
|
|
template class TQPtrList<KIO::Job>;
|
|
|
|
//this will update the report dialog with 5 Hz, I think this is fast enough, aleXXX
|
|
#define REPORT_TIMEOUT 200
|
|
|
|
#define KIO_ARGS TQByteArray packedArgs; TQDataStream stream( packedArgs, IO_WriteOnly ); stream
|
|
|
|
class Job::JobPrivate
|
|
{
|
|
public:
|
|
JobPrivate() : m_autoErrorHandling( false ), m_autoWarningHandling( true ),
|
|
m_interactive( true ), m_parentJob( 0L ), m_extraFlags(0),
|
|
m_processedSize(0), m_userTimestamp(0)
|
|
{}
|
|
|
|
bool m_autoErrorHandling;
|
|
bool m_autoWarningHandling;
|
|
bool m_interactive;
|
|
TQGuardedPtr<TQWidget> m_errorParentWidget;
|
|
// Maybe we could use the TQObject parent/child mechanism instead
|
|
// (requires a new ctor, and moving the ctor code to some init()).
|
|
Job* m_parentJob;
|
|
int m_extraFlags;
|
|
KIO::filesize_t m_processedSize;
|
|
unsigned long m_userTimestamp;
|
|
};
|
|
|
|
Job::Job(bool showProgressInfo) : TQObject(0, "job"), m_error(0), m_percent(0)
|
|
, m_progressId(0), m_speedTimer(0), d( new JobPrivate )
|
|
{
|
|
// All jobs delete themselves after emiting 'result'.
|
|
|
|
// Notify the UI Server and get a progress id
|
|
if ( showProgressInfo )
|
|
{
|
|
m_progressId = Observer::self()->newJob( this, true );
|
|
addMetaData("progress-id", TQString::number(m_progressId));
|
|
//kdDebug(7007) << "Created job " << this << " with progress info -- m_progressId=" << m_progressId << endl;
|
|
// Connect global progress info signals
|
|
connect( this, TQT_SIGNAL( percent( KIO::Job*, unsigned long ) ),
|
|
Observer::self(), TQT_SLOT( slotPercent( KIO::Job*, unsigned long ) ) );
|
|
connect( this, TQT_SIGNAL( infoMessage( KIO::Job*, const TQString & ) ),
|
|
Observer::self(), TQT_SLOT( slotInfoMessage( KIO::Job*, const TQString & ) ) );
|
|
connect( this, TQT_SIGNAL( totalSize( KIO::Job*, KIO::filesize_t ) ),
|
|
Observer::self(), TQT_SLOT( slotTotalSize( KIO::Job*, KIO::filesize_t ) ) );
|
|
connect( this, TQT_SIGNAL( processedSize( KIO::Job*, KIO::filesize_t ) ),
|
|
Observer::self(), TQT_SLOT( slotProcessedSize( KIO::Job*, KIO::filesize_t ) ) );
|
|
connect( this, TQT_SIGNAL( speed( KIO::Job*, unsigned long ) ),
|
|
Observer::self(), TQT_SLOT( slotSpeed( KIO::Job*, unsigned long ) ) );
|
|
}
|
|
// Don't exit while this job is running
|
|
if (kapp)
|
|
kapp->ref();
|
|
if (kapp)
|
|
updateUserTimestamp( kapp->userTimestamp());
|
|
}
|
|
|
|
Job::~Job()
|
|
{
|
|
delete m_speedTimer;
|
|
delete d;
|
|
if (kapp)
|
|
kapp->deref();
|
|
}
|
|
|
|
int& Job::extraFlags()
|
|
{
|
|
return d->m_extraFlags;
|
|
}
|
|
|
|
void Job::setProcessedSize(KIO::filesize_t size)
|
|
{
|
|
d->m_processedSize = size;
|
|
}
|
|
|
|
KIO::filesize_t Job::getProcessedSize()
|
|
{
|
|
return d->m_processedSize;
|
|
}
|
|
|
|
void Job::addSubjob(Job *job, bool inheritMetaData)
|
|
{
|
|
//kdDebug(7007) << "addSubjob(" << job << ") this = " << this << endl;
|
|
subjobs.append(job);
|
|
|
|
connect( job, TQT_SIGNAL(result(KIO::Job*)),
|
|
TQT_SLOT(slotResult(KIO::Job*)) );
|
|
|
|
// Forward information from that subjob.
|
|
connect( job, TQT_SIGNAL(speed( KIO::Job*, unsigned long )),
|
|
TQT_SLOT(slotSpeed(KIO::Job*, unsigned long)) );
|
|
|
|
connect( job, TQT_SIGNAL(infoMessage( KIO::Job*, const TQString & )),
|
|
TQT_SLOT(slotInfoMessage(KIO::Job*, const TQString &)) );
|
|
|
|
if (inheritMetaData)
|
|
job->mergeMetaData(m_outgoingMetaData);
|
|
|
|
job->setWindow( m_window );
|
|
job->updateUserTimestamp( d->m_userTimestamp );
|
|
}
|
|
|
|
void Job::removeSubjob( Job *job )
|
|
{
|
|
removeSubjob( job, false, true );
|
|
}
|
|
|
|
void Job::removeSubjob( Job *job, bool mergeMetaData, bool emitResultIfLast )
|
|
{
|
|
//kdDebug(7007) << "removeSubjob(" << job << ") this = " << this << " subjobs = " << subjobs.count() << endl;
|
|
// Merge metadata from subjob
|
|
if ( mergeMetaData )
|
|
m_incomingMetaData += job->metaData();
|
|
subjobs.remove(job);
|
|
if ( subjobs.isEmpty() && emitResultIfLast )
|
|
emitResult();
|
|
}
|
|
|
|
void Job::emitPercent( KIO::filesize_t processedSize, KIO::filesize_t totalSize )
|
|
{
|
|
// calculate percents
|
|
unsigned long ipercent = m_percent;
|
|
|
|
if ( totalSize == 0 )
|
|
m_percent = 100;
|
|
else
|
|
m_percent = (unsigned long)(( (float)(processedSize) / (float)(totalSize) ) * 100.0);
|
|
|
|
if ( m_percent != ipercent || m_percent == 100 /* for those buggy total sizes that grow */ ) {
|
|
emit percent( this, m_percent );
|
|
//kdDebug(7007) << "Job::emitPercent - percent = " << (unsigned int) m_percent << endl;
|
|
}
|
|
}
|
|
|
|
void Job::emitSpeed( unsigned long bytes_per_second )
|
|
{
|
|
//kdDebug(7007) << "Job " << this << " emitSpeed " << bytes_per_second << endl;
|
|
if ( !m_speedTimer )
|
|
{
|
|
m_speedTimer = new TQTimer();
|
|
connect( m_speedTimer, TQT_SIGNAL( timeout() ), TQT_SLOT( slotSpeedTimeout() ) );
|
|
}
|
|
emit speed( this, bytes_per_second );
|
|
m_speedTimer->start( 5000 ); // 5 seconds interval should be enough
|
|
}
|
|
|
|
void Job::emitResult()
|
|
{
|
|
// If we are displaying a progress dialog, remove it first.
|
|
if ( m_progressId ) // Did we get an ID from the observer ?
|
|
Observer::self()->jobFinished( m_progressId );
|
|
if ( m_error && d->m_interactive && d->m_autoErrorHandling )
|
|
showErrorDialog( d->m_errorParentWidget );
|
|
emit result(this);
|
|
deleteLater();
|
|
}
|
|
|
|
void Job::kill( bool quietly )
|
|
{
|
|
kdDebug(7007) << "Job::kill this=" << this << " " << className() << " m_progressId=" << m_progressId << " quietly=" << quietly << endl;
|
|
// kill all subjobs, without triggering their result slot
|
|
TQPtrListIterator<Job> it( subjobs );
|
|
for ( ; it.current() ; ++it )
|
|
(*it)->kill( true );
|
|
subjobs.clear();
|
|
|
|
if ( ! quietly ) {
|
|
m_error = ERR_USER_CANCELED;
|
|
emit canceled( this ); // Not very useful (deprecated)
|
|
emitResult();
|
|
} else
|
|
{
|
|
if ( m_progressId ) // in both cases we want to hide the progress window
|
|
Observer::self()->jobFinished( m_progressId );
|
|
deleteLater();
|
|
}
|
|
}
|
|
|
|
void Job::slotResult( Job *job )
|
|
{
|
|
// Did job have an error ?
|
|
if ( job->error() && !m_error )
|
|
{
|
|
// Store it in the parent only if first error
|
|
m_error = job->error();
|
|
m_errorText = job->errorText();
|
|
}
|
|
removeSubjob(job);
|
|
}
|
|
|
|
void Job::slotSpeed( KIO::Job*, unsigned long speed )
|
|
{
|
|
//kdDebug(7007) << "Job::slotSpeed " << speed << endl;
|
|
emitSpeed( speed );
|
|
}
|
|
|
|
void Job::slotInfoMessage( KIO::Job*, const TQString & msg )
|
|
{
|
|
emit infoMessage( this, msg );
|
|
}
|
|
|
|
void Job::slotSpeedTimeout()
|
|
{
|
|
//kdDebug(7007) << "slotSpeedTimeout()" << endl;
|
|
// send 0 and stop the timer
|
|
// timer will be restarted only when we receive another speed event
|
|
emit speed( this, 0 );
|
|
m_speedTimer->stop();
|
|
}
|
|
|
|
//Job::errorString is implemented in global.cpp
|
|
|
|
void Job::showErrorDialog( TQWidget * parent )
|
|
{
|
|
//kdDebug(7007) << "Job::showErrorDialog parent=" << parent << endl;
|
|
kapp->enableStyles();
|
|
// Show a message box, except for "user canceled" or "no content"
|
|
if ( (m_error != ERR_USER_CANCELED) && (m_error != ERR_NO_CONTENT) ) {
|
|
//old plain error message
|
|
//kdDebug(7007) << "Default language: " << KGlobal::locale()->defaultLanguage() << endl;
|
|
if ( 1 )
|
|
KMessageBox::queuedMessageBox( parent, KMessageBox::Error, errorString() );
|
|
#if 0
|
|
} else {
|
|
TQStringList errors = detailedErrorStrings();
|
|
TQString caption, err, detail;
|
|
TQStringList::const_iterator it = errors.begin();
|
|
if ( it != errors.end() )
|
|
caption = *(it++);
|
|
if ( it != errors.end() )
|
|
err = *(it++);
|
|
if ( it != errors.end() )
|
|
detail = *it;
|
|
KMessageBox::queuedDetailedError( parent, err, detail, caption );
|
|
}
|
|
#endif
|
|
}
|
|
}
|
|
|
|
void Job::setAutoErrorHandlingEnabled( bool enable, TQWidget *parentWidget )
|
|
{
|
|
d->m_autoErrorHandling = enable;
|
|
d->m_errorParentWidget = parentWidget;
|
|
}
|
|
|
|
bool Job::isAutoErrorHandlingEnabled() const
|
|
{
|
|
return d->m_autoErrorHandling;
|
|
}
|
|
|
|
void Job::setAutoWarningHandlingEnabled( bool enable )
|
|
{
|
|
d->m_autoWarningHandling = enable;
|
|
}
|
|
|
|
bool Job::isAutoWarningHandlingEnabled() const
|
|
{
|
|
return d->m_autoWarningHandling;
|
|
}
|
|
|
|
void Job::setInteractive(bool enable)
|
|
{
|
|
d->m_interactive = enable;
|
|
}
|
|
|
|
bool Job::isInteractive() const
|
|
{
|
|
return d->m_interactive;
|
|
}
|
|
|
|
void Job::setWindow(TQWidget *window)
|
|
{
|
|
m_window = window;
|
|
KIO::Scheduler::registerWindow(window);
|
|
}
|
|
|
|
TQWidget *Job::window() const
|
|
{
|
|
return m_window;
|
|
}
|
|
|
|
void Job::updateUserTimestamp( unsigned long time )
|
|
{
|
|
#if defined Q_WS_X11
|
|
if( d->m_userTimestamp == 0 || NET::timestampCompare( time, d->m_userTimestamp ) > 0 )
|
|
d->m_userTimestamp = time;
|
|
#endif
|
|
}
|
|
|
|
unsigned long Job::userTimestamp() const
|
|
{
|
|
return d->m_userTimestamp;
|
|
}
|
|
|
|
void Job::setParentJob(Job* job)
|
|
{
|
|
Q_ASSERT(d->m_parentJob == 0L);
|
|
Q_ASSERT(job);
|
|
d->m_parentJob = job;
|
|
}
|
|
|
|
Job* Job::parentJob() const
|
|
{
|
|
return d->m_parentJob;
|
|
}
|
|
|
|
MetaData Job::metaData() const
|
|
{
|
|
return m_incomingMetaData;
|
|
}
|
|
|
|
TQString Job::queryMetaData(const TQString &key)
|
|
{
|
|
if (!m_incomingMetaData.contains(key))
|
|
return TQString::null;
|
|
return m_incomingMetaData[key];
|
|
}
|
|
|
|
void Job::setMetaData( const KIO::MetaData &_metaData)
|
|
{
|
|
m_outgoingMetaData = _metaData;
|
|
}
|
|
|
|
void Job::addMetaData( const TQString &key, const TQString &value)
|
|
{
|
|
m_outgoingMetaData.insert(key, value);
|
|
}
|
|
|
|
void Job::addMetaData( const TQMap<TQString,TQString> &values)
|
|
{
|
|
TQMapConstIterator<TQString,TQString> it = values.begin();
|
|
for(;it != values.end(); ++it)
|
|
m_outgoingMetaData.insert(it.key(), it.data());
|
|
}
|
|
|
|
void Job::mergeMetaData( const TQMap<TQString,TQString> &values)
|
|
{
|
|
TQMapConstIterator<TQString,TQString> it = values.begin();
|
|
for(;it != values.end(); ++it)
|
|
m_outgoingMetaData.insert(it.key(), it.data(), false);
|
|
}
|
|
|
|
MetaData Job::outgoingMetaData() const
|
|
{
|
|
return m_outgoingMetaData;
|
|
}
|
|
|
|
|
|
SimpleJob::SimpleJob(const KURL& url, int command, const TQByteArray &packedArgs,
|
|
bool showProgressInfo )
|
|
: Job(showProgressInfo), m_slave(0), m_packedArgs(packedArgs),
|
|
m_url(url), m_command(command), m_totalSize(0)
|
|
{
|
|
if (m_url.hasSubURL())
|
|
{
|
|
KURL::List list = KURL::split(m_url);
|
|
KURL::List::Iterator it = list.fromLast();
|
|
list.remove(it);
|
|
m_subUrl = KURL::join(list);
|
|
//kdDebug(7007) << "New URL = " << m_url.url() << endl;
|
|
//kdDebug(7007) << "Sub URL = " << m_subUrl.url() << endl;
|
|
}
|
|
|
|
Scheduler::doJob(this);
|
|
|
|
if (!m_url.isValid())
|
|
{
|
|
kdDebug() << "ERR_MALFORMED_URL" << endl;
|
|
m_error = ERR_MALFORMED_URL;
|
|
m_errorText = m_url.url();
|
|
TQTimer::singleShot(0, this, TQT_SLOT(slotFinished()) );
|
|
return;
|
|
}
|
|
}
|
|
|
|
void SimpleJob::kill( bool quietly )
|
|
{
|
|
Scheduler::cancelJob( this ); // deletes the slave if not 0
|
|
m_slave = 0; // -> set to 0
|
|
Job::kill( quietly );
|
|
}
|
|
|
|
void SimpleJob::putOnHold()
|
|
{
|
|
Q_ASSERT( m_slave );
|
|
if ( m_slave )
|
|
{
|
|
Scheduler::putSlaveOnHold(this, m_url);
|
|
m_slave = 0;
|
|
}
|
|
kill(true);
|
|
}
|
|
|
|
void SimpleJob::removeOnHold()
|
|
{
|
|
Scheduler::removeSlaveOnHold();
|
|
}
|
|
|
|
SimpleJob::~SimpleJob()
|
|
{
|
|
if (m_slave) // was running
|
|
{
|
|
kdDebug(7007) << "SimpleJob::~SimpleJob: Killing running job in destructor!" << endl;
|
|
#if 0
|
|
m_slave->kill();
|
|
Scheduler::jobFinished( this, m_slave ); // deletes the slave
|
|
#endif
|
|
Scheduler::cancelJob( this );
|
|
m_slave = 0; // -> set to 0
|
|
}
|
|
}
|
|
|
|
void SimpleJob::start(Slave *slave)
|
|
{
|
|
m_slave = slave;
|
|
|
|
connect( m_slave, TQT_SIGNAL( error( int , const TQString & ) ),
|
|
TQT_SLOT( slotError( int , const TQString & ) ) );
|
|
|
|
connect( m_slave, TQT_SIGNAL( warning( const TQString & ) ),
|
|
TQT_SLOT( slotWarning( const TQString & ) ) );
|
|
|
|
connect( m_slave, TQT_SIGNAL( infoMessage( const TQString & ) ),
|
|
TQT_SLOT( slotInfoMessage( const TQString & ) ) );
|
|
|
|
connect( m_slave, TQT_SIGNAL( connected() ),
|
|
TQT_SLOT( slotConnected() ) );
|
|
|
|
connect( m_slave, TQT_SIGNAL( finished() ),
|
|
TQT_SLOT( slotFinished() ) );
|
|
|
|
if ((extraFlags() & EF_TransferJobDataSent) == 0)
|
|
{
|
|
connect( m_slave, TQT_SIGNAL( totalSize( KIO::filesize_t ) ),
|
|
TQT_SLOT( slotTotalSize( KIO::filesize_t ) ) );
|
|
|
|
connect( m_slave, TQT_SIGNAL( processedSize( KIO::filesize_t ) ),
|
|
TQT_SLOT( slotProcessedSize( KIO::filesize_t ) ) );
|
|
|
|
connect( m_slave, TQT_SIGNAL( speed( unsigned long ) ),
|
|
TQT_SLOT( slotSpeed( unsigned long ) ) );
|
|
}
|
|
|
|
connect( slave, TQT_SIGNAL( needProgressId() ),
|
|
TQT_SLOT( slotNeedProgressId() ) );
|
|
|
|
connect( slave, TQT_SIGNAL(metaData( const KIO::MetaData& ) ),
|
|
TQT_SLOT( slotMetaData( const KIO::MetaData& ) ) );
|
|
|
|
if (m_window)
|
|
{
|
|
TQString id;
|
|
addMetaData("window-id", id.setNum((ulong)m_window->winId()));
|
|
}
|
|
if (userTimestamp())
|
|
{
|
|
TQString id;
|
|
addMetaData("user-timestamp", id.setNum(userTimestamp()));
|
|
}
|
|
|
|
TQString sslSession = KSSLCSessionCache::getSessionForURL(m_url);
|
|
if ( !sslSession.isNull() )
|
|
{
|
|
addMetaData("ssl_session_id", sslSession);
|
|
}
|
|
|
|
if (!isInteractive())
|
|
{
|
|
addMetaData("no-auth-prompt", "true");
|
|
}
|
|
|
|
if (!m_outgoingMetaData.isEmpty())
|
|
{
|
|
KIO_ARGS << m_outgoingMetaData;
|
|
slave->send( CMD_META_DATA, packedArgs );
|
|
}
|
|
|
|
if (!m_subUrl.isEmpty())
|
|
{
|
|
KIO_ARGS << m_subUrl;
|
|
m_slave->send( CMD_SUBURL, packedArgs );
|
|
}
|
|
|
|
m_slave->send( m_command, m_packedArgs );
|
|
}
|
|
|
|
void SimpleJob::slaveDone()
|
|
{
|
|
if (!m_slave) return;
|
|
disconnect(m_slave); // Remove all signals between slave and job
|
|
Scheduler::jobFinished( this, m_slave );
|
|
m_slave = 0;
|
|
}
|
|
|
|
void SimpleJob::slotFinished( )
|
|
{
|
|
// Return slave to the scheduler
|
|
slaveDone();
|
|
|
|
if (subjobs.isEmpty())
|
|
{
|
|
if ( !m_error && (m_command == CMD_MKDIR || m_command == CMD_RENAME ) )
|
|
{
|
|
KDirNotify_stub allDirNotify( "*", "KDirNotify*" );
|
|
if ( m_command == CMD_MKDIR )
|
|
{
|
|
KURL urlDir( url() );
|
|
urlDir.setPath( urlDir.directory() );
|
|
allDirNotify.FilesAdded( urlDir );
|
|
}
|
|
else /*if ( m_command == CMD_RENAME )*/
|
|
{
|
|
KURL src, dst;
|
|
TQDataStream str( m_packedArgs, IO_ReadOnly );
|
|
str >> src >> dst;
|
|
if ( src.directory() == dst.directory() ) // For the user, moving isn't renaming. Only renaming is.
|
|
allDirNotify.FileRenamed( src, dst );
|
|
}
|
|
}
|
|
emitResult();
|
|
}
|
|
}
|
|
|
|
void SimpleJob::slotError( int error, const TQString & errorText )
|
|
{
|
|
m_error = error;
|
|
m_errorText = errorText;
|
|
if ((m_error == ERR_UNKNOWN_HOST) && m_url.host().isEmpty())
|
|
m_errorText = TQString::null;
|
|
// error terminates the job
|
|
slotFinished();
|
|
}
|
|
|
|
void SimpleJob::slotWarning( const TQString & errorText )
|
|
{
|
|
TQGuardedPtr<SimpleJob> guard( this );
|
|
if (isInteractive() && isAutoWarningHandlingEnabled())
|
|
{
|
|
static uint msgBoxDisplayed = 0;
|
|
if ( msgBoxDisplayed == 0 ) // don't bomb the user with message boxes, only one at a time
|
|
{
|
|
msgBoxDisplayed++;
|
|
KMessageBox::information( 0L, errorText );
|
|
msgBoxDisplayed--;
|
|
}
|
|
// otherwise just discard it.
|
|
}
|
|
|
|
if ( !guard.isNull() )
|
|
emit warning( this, errorText );
|
|
}
|
|
|
|
void SimpleJob::slotInfoMessage( const TQString & msg )
|
|
{
|
|
emit infoMessage( this, msg );
|
|
}
|
|
|
|
void SimpleJob::slotConnected()
|
|
{
|
|
emit connected( this );
|
|
}
|
|
|
|
void SimpleJob::slotNeedProgressId()
|
|
{
|
|
if ( !m_progressId )
|
|
m_progressId = Observer::self()->newJob( this, false );
|
|
m_slave->setProgressId( m_progressId );
|
|
}
|
|
|
|
void SimpleJob::slotTotalSize( KIO::filesize_t size )
|
|
{
|
|
if (size > m_totalSize)
|
|
{
|
|
m_totalSize = size;
|
|
emit totalSize( this, size );
|
|
}
|
|
}
|
|
|
|
void SimpleJob::slotProcessedSize( KIO::filesize_t size )
|
|
{
|
|
//kdDebug(7007) << "SimpleJob::slotProcessedSize " << KIO::number(size) << endl;
|
|
setProcessedSize(size);
|
|
emit processedSize( this, size );
|
|
if ( size > m_totalSize ) {
|
|
slotTotalSize(size); // safety
|
|
}
|
|
emitPercent( size, m_totalSize );
|
|
}
|
|
|
|
void SimpleJob::slotSpeed( unsigned long speed )
|
|
{
|
|
//kdDebug(7007) << "SimpleJob::slotSpeed( " << speed << " )" << endl;
|
|
emitSpeed( speed );
|
|
}
|
|
|
|
void SimpleJob::slotMetaData( const KIO::MetaData &_metaData)
|
|
{
|
|
m_incomingMetaData += _metaData;
|
|
}
|
|
|
|
void SimpleJob::storeSSLSessionFromJob(const KURL &m_redirectionURL) {
|
|
TQString sslSession = queryMetaData("ssl_session_id");
|
|
|
|
if ( !sslSession.isNull() ) {
|
|
const KURL &queryURL = m_redirectionURL.isEmpty()?m_url:m_redirectionURL;
|
|
KSSLCSessionCache::putSessionForURL(queryURL, sslSession);
|
|
}
|
|
}
|
|
|
|
//////////
|
|
MkdirJob::MkdirJob( const KURL& url, int command,
|
|
const TQByteArray &packedArgs, bool showProgressInfo )
|
|
: SimpleJob(url, command, packedArgs, showProgressInfo)
|
|
{
|
|
}
|
|
|
|
void MkdirJob::start(Slave *slave)
|
|
{
|
|
connect( slave, TQT_SIGNAL( redirection(const KURL &) ),
|
|
TQT_SLOT( slotRedirection(const KURL &) ) );
|
|
|
|
SimpleJob::start(slave);
|
|
}
|
|
|
|
// Slave got a redirection request
|
|
void MkdirJob::slotRedirection( const KURL &url)
|
|
{
|
|
kdDebug(7007) << "MkdirJob::slotRedirection(" << url << ")" << endl;
|
|
if (!kapp->authorizeURLAction("redirect", m_url, url))
|
|
{
|
|
kdWarning(7007) << "MkdirJob: Redirection from " << m_url << " to " << url << " REJECTED!" << endl;
|
|
m_error = ERR_ACCESS_DENIED;
|
|
m_errorText = url.prettyURL();
|
|
return;
|
|
}
|
|
m_redirectionURL = url; // We'll remember that when the job finishes
|
|
if (m_url.hasUser() && !url.hasUser() && (m_url.host().lower() == url.host().lower()))
|
|
m_redirectionURL.setUser(m_url.user()); // Preserve user
|
|
// Tell the user that we haven't finished yet
|
|
emit redirection(this, m_redirectionURL);
|
|
}
|
|
|
|
void MkdirJob::slotFinished()
|
|
{
|
|
if ( m_redirectionURL.isEmpty() || !m_redirectionURL.isValid())
|
|
{
|
|
// Return slave to the scheduler
|
|
SimpleJob::slotFinished();
|
|
} else {
|
|
//kdDebug(7007) << "MkdirJob: Redirection to " << m_redirectionURL << endl;
|
|
if (queryMetaData("permanent-redirect")=="true")
|
|
emit permanentRedirection(this, m_url, m_redirectionURL);
|
|
KURL dummyUrl;
|
|
int permissions;
|
|
TQDataStream istream( m_packedArgs, IO_ReadOnly );
|
|
istream >> dummyUrl >> permissions;
|
|
|
|
m_url = m_redirectionURL;
|
|
m_redirectionURL = KURL();
|
|
m_packedArgs.truncate(0);
|
|
TQDataStream stream( m_packedArgs, IO_WriteOnly );
|
|
stream << m_url << permissions;
|
|
|
|
// Return slave to the scheduler
|
|
slaveDone();
|
|
Scheduler::doJob(this);
|
|
}
|
|
}
|
|
|
|
SimpleJob *KIO::mkdir( const KURL& url, int permissions )
|
|
{
|
|
//kdDebug(7007) << "mkdir " << url << endl;
|
|
KIO_ARGS << url << permissions;
|
|
return new MkdirJob(url, CMD_MKDIR, packedArgs, false);
|
|
}
|
|
|
|
SimpleJob *KIO::rmdir( const KURL& url )
|
|
{
|
|
//kdDebug(7007) << "rmdir " << url << endl;
|
|
KIO_ARGS << url << TQ_INT8(false); // isFile is false
|
|
return new SimpleJob(url, CMD_DEL, packedArgs, false);
|
|
}
|
|
|
|
SimpleJob *KIO::chmod( const KURL& url, int permissions )
|
|
{
|
|
//kdDebug(7007) << "chmod " << url << endl;
|
|
KIO_ARGS << url << permissions;
|
|
return new SimpleJob(url, CMD_CHMOD, packedArgs, false);
|
|
}
|
|
|
|
SimpleJob *KIO::rename( const KURL& src, const KURL & dest, bool overwrite )
|
|
{
|
|
//kdDebug(7007) << "rename " << src << " " << dest << endl;
|
|
KIO_ARGS << src << dest << (TQ_INT8) overwrite;
|
|
return new SimpleJob(src, CMD_RENAME, packedArgs, false);
|
|
}
|
|
|
|
SimpleJob *KIO::symlink( const TQString& target, const KURL & dest, bool overwrite, bool showProgressInfo )
|
|
{
|
|
//kdDebug(7007) << "symlink target=" << target << " " << dest << endl;
|
|
KIO_ARGS << target << dest << (TQ_INT8) overwrite;
|
|
return new SimpleJob(dest, CMD_SYMLINK, packedArgs, showProgressInfo);
|
|
}
|
|
|
|
SimpleJob *KIO::special(const KURL& url, const TQByteArray & data, bool showProgressInfo)
|
|
{
|
|
//kdDebug(7007) << "special " << url << endl;
|
|
return new SimpleJob(url, CMD_SPECIAL, data, showProgressInfo);
|
|
}
|
|
|
|
SimpleJob *KIO::mount( bool ro, const char *fstype, const TQString& dev, const TQString& point, bool showProgressInfo )
|
|
{
|
|
KIO_ARGS << int(1) << TQ_INT8( ro ? 1 : 0 )
|
|
<< TQString::tqfromLatin1(fstype) << dev << point;
|
|
SimpleJob *job = special( KURL("file:/"), packedArgs, showProgressInfo );
|
|
if ( showProgressInfo )
|
|
Observer::self()->mounting( job, dev, point );
|
|
return job;
|
|
}
|
|
|
|
SimpleJob *KIO::unmount( const TQString& point, bool showProgressInfo )
|
|
{
|
|
KIO_ARGS << int(2) << point;
|
|
SimpleJob *job = special( KURL("file:/"), packedArgs, showProgressInfo );
|
|
if ( showProgressInfo )
|
|
Observer::self()->unmounting( job, point );
|
|
return job;
|
|
}
|
|
|
|
|
|
|
|
//////////
|
|
|
|
StatJob::StatJob( const KURL& url, int command,
|
|
const TQByteArray &packedArgs, bool showProgressInfo )
|
|
: SimpleJob(url, command, packedArgs, showProgressInfo),
|
|
m_bSource(true), m_details(2)
|
|
{
|
|
}
|
|
|
|
void StatJob::start(Slave *slave)
|
|
{
|
|
m_outgoingMetaData.replace( "statSide", m_bSource ? "source" : "dest" );
|
|
m_outgoingMetaData.replace( "details", TQString::number(m_details) );
|
|
|
|
connect( slave, TQT_SIGNAL( statEntry( const KIO::UDSEntry& ) ),
|
|
TQT_SLOT( slotStatEntry( const KIO::UDSEntry & ) ) );
|
|
connect( slave, TQT_SIGNAL( redirection(const KURL &) ),
|
|
TQT_SLOT( slotRedirection(const KURL &) ) );
|
|
|
|
SimpleJob::start(slave);
|
|
}
|
|
|
|
void StatJob::slotStatEntry( const KIO::UDSEntry & entry )
|
|
{
|
|
//kdDebug(7007) << "StatJob::slotStatEntry" << endl;
|
|
m_statResult = entry;
|
|
}
|
|
|
|
// Slave got a redirection request
|
|
void StatJob::slotRedirection( const KURL &url)
|
|
{
|
|
kdDebug(7007) << "StatJob::slotRedirection(" << url << ")" << endl;
|
|
if (!kapp->authorizeURLAction("redirect", m_url, url))
|
|
{
|
|
kdWarning(7007) << "StatJob: Redirection from " << m_url << " to " << url << " REJECTED!" << endl;
|
|
m_error = ERR_ACCESS_DENIED;
|
|
m_errorText = url.prettyURL();
|
|
return;
|
|
}
|
|
m_redirectionURL = url; // We'll remember that when the job finishes
|
|
if (m_url.hasUser() && !url.hasUser() && (m_url.host().lower() == url.host().lower()))
|
|
m_redirectionURL.setUser(m_url.user()); // Preserve user
|
|
// Tell the user that we haven't finished yet
|
|
emit redirection(this, m_redirectionURL);
|
|
}
|
|
|
|
void StatJob::slotFinished()
|
|
{
|
|
if ( m_redirectionURL.isEmpty() || !m_redirectionURL.isValid())
|
|
{
|
|
// Return slave to the scheduler
|
|
SimpleJob::slotFinished();
|
|
} else {
|
|
//kdDebug(7007) << "StatJob: Redirection to " << m_redirectionURL << endl;
|
|
if (queryMetaData("permanent-redirect")=="true")
|
|
emit permanentRedirection(this, m_url, m_redirectionURL);
|
|
m_url = m_redirectionURL;
|
|
m_redirectionURL = KURL();
|
|
m_packedArgs.truncate(0);
|
|
TQDataStream stream( m_packedArgs, IO_WriteOnly );
|
|
stream << m_url;
|
|
|
|
// Return slave to the scheduler
|
|
slaveDone();
|
|
Scheduler::doJob(this);
|
|
}
|
|
}
|
|
|
|
void StatJob::slotMetaData( const KIO::MetaData &_metaData) {
|
|
SimpleJob::slotMetaData(_metaData);
|
|
storeSSLSessionFromJob(m_redirectionURL);
|
|
}
|
|
|
|
StatJob *KIO::stat(const KURL& url, bool showProgressInfo)
|
|
{
|
|
// Assume sideIsSource. Gets are more common than puts.
|
|
return stat( url, true, 2, showProgressInfo );
|
|
}
|
|
|
|
StatJob *KIO::stat(const KURL& url, bool sideIsSource, short int details, bool showProgressInfo)
|
|
{
|
|
kdDebug(7007) << "stat " << url << endl;
|
|
KIO_ARGS << url;
|
|
StatJob * job = new StatJob(url, CMD_STAT, packedArgs, showProgressInfo );
|
|
job->setSide( sideIsSource );
|
|
job->setDetails( details );
|
|
if ( showProgressInfo )
|
|
Observer::self()->stating( job, url );
|
|
return job;
|
|
}
|
|
|
|
SimpleJob *KIO::http_update_cache( const KURL& url, bool no_cache, time_t expireDate)
|
|
{
|
|
assert( (url.protocol() == "http") || (url.protocol() == "https") );
|
|
// Send http update_cache command (2)
|
|
KIO_ARGS << (int)2 << url << no_cache << expireDate;
|
|
SimpleJob * job = new SimpleJob( url, CMD_SPECIAL, packedArgs, false );
|
|
Scheduler::scheduleJob(job);
|
|
return job;
|
|
}
|
|
|
|
//////////
|
|
|
|
TransferJob::TransferJob( const KURL& url, int command,
|
|
const TQByteArray &packedArgs,
|
|
const TQByteArray &_staticData,
|
|
bool showProgressInfo)
|
|
: SimpleJob(url, command, packedArgs, showProgressInfo), staticData( _staticData)
|
|
{
|
|
m_suspended = false;
|
|
m_errorPage = false;
|
|
m_subJob = 0L;
|
|
if ( showProgressInfo )
|
|
Observer::self()->slotTransferring( this, url );
|
|
}
|
|
|
|
// Slave sends data
|
|
void TransferJob::slotData( const TQByteArray &_data)
|
|
{
|
|
if(m_redirectionURL.isEmpty() || !m_redirectionURL.isValid() || m_error)
|
|
emit data( this, _data);
|
|
}
|
|
|
|
// Slave got a redirection request
|
|
void TransferJob::slotRedirection( const KURL &url)
|
|
{
|
|
kdDebug(7007) << "TransferJob::slotRedirection(" << url << ")" << endl;
|
|
if (!kapp->authorizeURLAction("redirect", m_url, url))
|
|
{
|
|
kdWarning(7007) << "TransferJob: Redirection from " << m_url << " to " << url << " REJECTED!" << endl;
|
|
return;
|
|
}
|
|
|
|
// Some websites keep redirecting to themselves where each redirection
|
|
// acts as the stage in a state-machine. We define "endless redirections"
|
|
// as 5 redirections to the same URL.
|
|
if (m_redirectionList.contains(url) > 5)
|
|
{
|
|
kdDebug(7007) << "TransferJob::slotRedirection: CYCLIC REDIRECTION!" << endl;
|
|
m_error = ERR_CYCLIC_LINK;
|
|
m_errorText = m_url.prettyURL();
|
|
}
|
|
else
|
|
{
|
|
m_redirectionURL = url; // We'll remember that when the job finishes
|
|
if (m_url.hasUser() && !url.hasUser() && (m_url.host().lower() == url.host().lower()))
|
|
m_redirectionURL.setUser(m_url.user()); // Preserve user
|
|
m_redirectionList.append(url);
|
|
m_outgoingMetaData["ssl_was_in_use"] = m_incomingMetaData["ssl_in_use"];
|
|
// Tell the user that we haven't finished yet
|
|
emit redirection(this, m_redirectionURL);
|
|
}
|
|
}
|
|
|
|
void TransferJob::slotFinished()
|
|
{
|
|
//kdDebug(7007) << "TransferJob::slotFinished(" << this << ", " << m_url << ")" << endl;
|
|
if (m_redirectionURL.isEmpty() || !m_redirectionURL.isValid())
|
|
SimpleJob::slotFinished();
|
|
else {
|
|
//kdDebug(7007) << "TransferJob: Redirection to " << m_redirectionURL << endl;
|
|
if (queryMetaData("permanent-redirect")=="true")
|
|
emit permanentRedirection(this, m_url, m_redirectionURL);
|
|
// Honour the redirection
|
|
// We take the approach of "redirecting this same job"
|
|
// Another solution would be to create a subjob, but the same problem
|
|
// happens (unpacking+repacking)
|
|
staticData.truncate(0);
|
|
m_incomingMetaData.clear();
|
|
if (queryMetaData("cache") != "reload")
|
|
addMetaData("cache","refresh");
|
|
m_suspended = false;
|
|
m_url = m_redirectionURL;
|
|
m_redirectionURL = KURL();
|
|
// The very tricky part is the packed arguments business
|
|
TQString dummyStr;
|
|
KURL dummyUrl;
|
|
TQDataStream istream( m_packedArgs, IO_ReadOnly );
|
|
switch( m_command ) {
|
|
case CMD_GET: {
|
|
m_packedArgs.truncate(0);
|
|
TQDataStream stream( m_packedArgs, IO_WriteOnly );
|
|
stream << m_url;
|
|
break;
|
|
}
|
|
case CMD_PUT: {
|
|
int permissions;
|
|
TQ_INT8 iOverwrite, iResume;
|
|
istream >> dummyUrl >> iOverwrite >> iResume >> permissions;
|
|
m_packedArgs.truncate(0);
|
|
TQDataStream stream( m_packedArgs, IO_WriteOnly );
|
|
stream << m_url << iOverwrite << iResume << permissions;
|
|
break;
|
|
}
|
|
case CMD_SPECIAL: {
|
|
int specialcmd;
|
|
istream >> specialcmd;
|
|
if (specialcmd == 1) // HTTP POST
|
|
{
|
|
addMetaData("cache","reload");
|
|
m_packedArgs.truncate(0);
|
|
TQDataStream stream( m_packedArgs, IO_WriteOnly );
|
|
stream << m_url;
|
|
m_command = CMD_GET;
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
|
|
// Return slave to the scheduler
|
|
slaveDone();
|
|
Scheduler::doJob(this);
|
|
}
|
|
}
|
|
|
|
void TransferJob::setAsyncDataEnabled(bool enabled)
|
|
{
|
|
if (enabled)
|
|
extraFlags() |= EF_TransferJobAsync;
|
|
else
|
|
extraFlags() &= ~EF_TransferJobAsync;
|
|
}
|
|
|
|
void TransferJob::sendAsyncData(const TQByteArray &dataForSlave)
|
|
{
|
|
if (extraFlags() & EF_TransferJobNeedData)
|
|
{
|
|
m_slave->send( MSG_DATA, dataForSlave );
|
|
if (extraFlags() & EF_TransferJobDataSent)
|
|
{
|
|
KIO::filesize_t size = getProcessedSize()+dataForSlave.size();
|
|
setProcessedSize(size);
|
|
emit processedSize( this, size );
|
|
if ( size > m_totalSize ) {
|
|
slotTotalSize(size); // safety
|
|
}
|
|
emitPercent( size, m_totalSize );
|
|
}
|
|
}
|
|
|
|
extraFlags() &= ~EF_TransferJobNeedData;
|
|
}
|
|
|
|
void TransferJob::setReportDataSent(bool enabled)
|
|
{
|
|
if (enabled)
|
|
extraFlags() |= EF_TransferJobDataSent;
|
|
else
|
|
extraFlags() &= ~EF_TransferJobDataSent;
|
|
}
|
|
|
|
bool TransferJob::reportDataSent()
|
|
{
|
|
return (extraFlags() & EF_TransferJobDataSent);
|
|
}
|
|
|
|
|
|
// Slave requests data
|
|
void TransferJob::slotDataReq()
|
|
{
|
|
TQByteArray dataForSlave;
|
|
|
|
extraFlags() |= EF_TransferJobNeedData;
|
|
|
|
if (!staticData.isEmpty())
|
|
{
|
|
dataForSlave = staticData;
|
|
staticData = TQByteArray();
|
|
}
|
|
else
|
|
{
|
|
emit dataReq( this, dataForSlave);
|
|
|
|
if (extraFlags() & EF_TransferJobAsync)
|
|
return;
|
|
}
|
|
|
|
static const size_t max_size = 14 * 1024 * 1024;
|
|
if (dataForSlave.size() > max_size)
|
|
{
|
|
kdDebug(7007) << "send " << dataForSlave.size() / 1024 / 1024 << "MB of data in TransferJob::dataReq. This needs to be splitted, which requires a copy. Fix the application.\n";
|
|
staticData.duplicate(dataForSlave.data() + max_size , dataForSlave.size() - max_size);
|
|
dataForSlave.truncate(max_size);
|
|
}
|
|
|
|
sendAsyncData(dataForSlave);
|
|
|
|
if (m_subJob)
|
|
{
|
|
// Bitburger protocol in action
|
|
suspend(); // Wait for more data from subJob.
|
|
m_subJob->resume(); // Ask for more!
|
|
}
|
|
}
|
|
|
|
void TransferJob::slotMimetype( const TQString& type )
|
|
{
|
|
m_mimetype = type;
|
|
emit mimetype( this, m_mimetype);
|
|
}
|
|
|
|
|
|
void TransferJob::suspend()
|
|
{
|
|
m_suspended = true;
|
|
if (m_slave)
|
|
m_slave->suspend();
|
|
}
|
|
|
|
void TransferJob::resume()
|
|
{
|
|
m_suspended = false;
|
|
if (m_slave)
|
|
m_slave->resume();
|
|
}
|
|
|
|
void TransferJob::start(Slave *slave)
|
|
{
|
|
assert(slave);
|
|
connect( slave, TQT_SIGNAL( data( const TQByteArray & ) ),
|
|
TQT_SLOT( slotData( const TQByteArray & ) ) );
|
|
|
|
connect( slave, TQT_SIGNAL( dataReq() ),
|
|
TQT_SLOT( slotDataReq() ) );
|
|
|
|
connect( slave, TQT_SIGNAL( redirection(const KURL &) ),
|
|
TQT_SLOT( slotRedirection(const KURL &) ) );
|
|
|
|
connect( slave, TQT_SIGNAL(mimeType( const TQString& ) ),
|
|
TQT_SLOT( slotMimetype( const TQString& ) ) );
|
|
|
|
connect( slave, TQT_SIGNAL(errorPage() ),
|
|
TQT_SLOT( slotErrorPage() ) );
|
|
|
|
connect( slave, TQT_SIGNAL( needSubURLData() ),
|
|
TQT_SLOT( slotNeedSubURLData() ) );
|
|
|
|
connect( slave, TQT_SIGNAL(canResume( KIO::filesize_t ) ),
|
|
TQT_SLOT( slotCanResume( KIO::filesize_t ) ) );
|
|
|
|
if (slave->suspended())
|
|
{
|
|
m_mimetype = "unknown";
|
|
// WABA: The slave was put on hold. Resume operation.
|
|
slave->resume();
|
|
}
|
|
|
|
SimpleJob::start(slave);
|
|
if (m_suspended)
|
|
slave->suspend();
|
|
}
|
|
|
|
void TransferJob::slotNeedSubURLData()
|
|
{
|
|
// Job needs data from subURL.
|
|
m_subJob = KIO::get( m_subUrl, false, false);
|
|
suspend(); // Put job on hold until we have some data.
|
|
connect(m_subJob, TQT_SIGNAL( data(KIO::Job*,const TQByteArray &)),
|
|
TQT_SLOT( slotSubURLData(KIO::Job*,const TQByteArray &)));
|
|
addSubjob(m_subJob);
|
|
}
|
|
|
|
void TransferJob::slotSubURLData(KIO::Job*, const TQByteArray &data)
|
|
{
|
|
// The Alternating Bitburg protocol in action again.
|
|
staticData = data;
|
|
m_subJob->suspend(); // Put job on hold until we have delivered the data.
|
|
resume(); // Activate ourselves again.
|
|
}
|
|
|
|
void TransferJob::slotMetaData( const KIO::MetaData &_metaData) {
|
|
SimpleJob::slotMetaData(_metaData);
|
|
storeSSLSessionFromJob(m_redirectionURL);
|
|
}
|
|
|
|
void TransferJob::slotErrorPage()
|
|
{
|
|
m_errorPage = true;
|
|
}
|
|
|
|
void TransferJob::slotCanResume( KIO::filesize_t offset )
|
|
{
|
|
emit canResume(this, offset);
|
|
}
|
|
|
|
void TransferJob::slotResult( KIO::Job *job)
|
|
{
|
|
// This can only be our suburl.
|
|
assert(job == m_subJob);
|
|
// Did job have an error ?
|
|
if ( job->error() )
|
|
{
|
|
m_error = job->error();
|
|
m_errorText = job->errorText();
|
|
|
|
emitResult();
|
|
return;
|
|
}
|
|
|
|
if (job == m_subJob)
|
|
{
|
|
m_subJob = 0; // No action required
|
|
resume(); // Make sure we get the remaining data.
|
|
}
|
|
removeSubjob( job, false, false ); // Remove job, but don't kill this job.
|
|
}
|
|
|
|
TransferJob *KIO::get( const KURL& url, bool reload, bool showProgressInfo )
|
|
{
|
|
// Send decoded path and encoded query
|
|
KIO_ARGS << url;
|
|
TransferJob * job = new TransferJob( url, CMD_GET, packedArgs, TQByteArray(), showProgressInfo );
|
|
if (reload)
|
|
job->addMetaData("cache", "reload");
|
|
return job;
|
|
}
|
|
|
|
class PostErrorJob : public TransferJob
|
|
{
|
|
public:
|
|
|
|
PostErrorJob(int _error, const TQString& url, const TQByteArray &packedArgs, const TQByteArray &postData, bool showProgressInfo)
|
|
: TransferJob(KURL(), CMD_SPECIAL, packedArgs, postData, showProgressInfo)
|
|
{
|
|
m_error = _error;
|
|
m_errorText = url;
|
|
}
|
|
|
|
};
|
|
|
|
TransferJob *KIO::http_post( const KURL& url, const TQByteArray &postData, bool showProgressInfo )
|
|
{
|
|
int _error = 0;
|
|
|
|
// filter out some malicious ports
|
|
static const int bad_ports[] = {
|
|
1, // tcpmux
|
|
7, // echo
|
|
9, // discard
|
|
11, // systat
|
|
13, // daytime
|
|
15, // netstat
|
|
17, // qotd
|
|
19, // chargen
|
|
20, // ftp-data
|
|
21, // ftp-cntl
|
|
22, // ssh
|
|
23, // telnet
|
|
25, // smtp
|
|
37, // time
|
|
42, // name
|
|
43, // nicname
|
|
53, // domain
|
|
77, // priv-rjs
|
|
79, // finger
|
|
87, // ttylink
|
|
95, // supdup
|
|
101, // hostriame
|
|
102, // iso-tsap
|
|
103, // gppitnp
|
|
104, // acr-nema
|
|
109, // pop2
|
|
110, // pop3
|
|
111, // sunrpc
|
|
113, // auth
|
|
115, // sftp
|
|
117, // uucp-path
|
|
119, // nntp
|
|
123, // NTP
|
|
135, // loc-srv / epmap
|
|
139, // netbios
|
|
143, // imap2
|
|
179, // BGP
|
|
389, // ldap
|
|
512, // print / exec
|
|
513, // login
|
|
514, // shell
|
|
515, // printer
|
|
526, // tempo
|
|
530, // courier
|
|
531, // Chat
|
|
532, // netnews
|
|
540, // uucp
|
|
556, // remotefs
|
|
587, // sendmail
|
|
601, //
|
|
989, // ftps data
|
|
990, // ftps
|
|
992, // telnets
|
|
993, // imap/SSL
|
|
995, // pop3/SSL
|
|
1080, // SOCKS
|
|
2049, // nfs
|
|
4045, // lockd
|
|
6000, // x11
|
|
6667, // irc
|
|
0};
|
|
for (int cnt=0; bad_ports[cnt]; ++cnt)
|
|
if (url.port() == bad_ports[cnt])
|
|
{
|
|
_error = KIO::ERR_POST_DENIED;
|
|
break;
|
|
}
|
|
|
|
if( _error )
|
|
{
|
|
static bool override_loaded = false;
|
|
static TQValueList< int >* overriden_ports = NULL;
|
|
if( !override_loaded )
|
|
{
|
|
KConfig cfg( "kio_httprc", true );
|
|
overriden_ports = new TQValueList< int >;
|
|
*overriden_ports = cfg.readIntListEntry( "OverriddenPorts" );
|
|
override_loaded = true;
|
|
}
|
|
for( TQValueList< int >::ConstIterator it = overriden_ports->begin();
|
|
it != overriden_ports->end();
|
|
++it )
|
|
if( overriden_ports->contains( url.port()))
|
|
_error = 0;
|
|
}
|
|
|
|
// filter out non https? protocols
|
|
if ((url.protocol() != "http") && (url.protocol() != "https" ))
|
|
_error = KIO::ERR_POST_DENIED;
|
|
|
|
bool redirection = false;
|
|
KURL _url(url);
|
|
if (_url.path().isEmpty())
|
|
{
|
|
redirection = true;
|
|
_url.setPath("/");
|
|
}
|
|
|
|
if (!_error && !kapp->authorizeURLAction("open", KURL(), _url))
|
|
_error = KIO::ERR_ACCESS_DENIED;
|
|
|
|
// if request is not valid, return an invalid transfer job
|
|
if (_error)
|
|
{
|
|
KIO_ARGS << (int)1 << url;
|
|
TransferJob * job = new PostErrorJob(_error, url.prettyURL(), packedArgs, postData, showProgressInfo);
|
|
return job;
|
|
}
|
|
|
|
// Send http post command (1), decoded path and encoded query
|
|
KIO_ARGS << (int)1 << _url;
|
|
TransferJob * job = new TransferJob( _url, CMD_SPECIAL,
|
|
packedArgs, postData, showProgressInfo );
|
|
|
|
if (redirection)
|
|
TQTimer::singleShot(0, job, TQT_SLOT(slotPostRedirection()) );
|
|
|
|
return job;
|
|
}
|
|
|
|
// http post got redirected from http://host to http://host/ by TransferJob
|
|
// We must do this redirection ourselves because redirections by the
|
|
// slave change post jobs into get jobs.
|
|
void TransferJob::slotPostRedirection()
|
|
{
|
|
kdDebug(7007) << "TransferJob::slotPostRedirection(" << m_url << ")" << endl;
|
|
// Tell the user about the new url.
|
|
emit redirection(this, m_url);
|
|
}
|
|
|
|
|
|
TransferJob *KIO::put( const KURL& url, int permissions,
|
|
bool overwrite, bool resume, bool showProgressInfo )
|
|
{
|
|
KIO_ARGS << url << TQ_INT8( overwrite ? 1 : 0 ) << TQ_INT8( resume ? 1 : 0 ) << permissions;
|
|
TransferJob * job = new TransferJob( url, CMD_PUT, packedArgs, TQByteArray(), showProgressInfo );
|
|
return job;
|
|
}
|
|
|
|
//////////
|
|
|
|
StoredTransferJob::StoredTransferJob(const KURL& url, int command,
|
|
const TQByteArray &packedArgs,
|
|
const TQByteArray &_staticData,
|
|
bool showProgressInfo)
|
|
: TransferJob( url, command, packedArgs, _staticData, showProgressInfo ),
|
|
m_uploadOffset( 0 )
|
|
{
|
|
connect( this, TQT_SIGNAL( data( KIO::Job *, const TQByteArray & ) ),
|
|
TQT_SLOT( slotStoredData( KIO::Job *, const TQByteArray & ) ) );
|
|
connect( this, TQT_SIGNAL( dataReq( KIO::Job *, TQByteArray & ) ),
|
|
TQT_SLOT( slotStoredDataReq( KIO::Job *, TQByteArray & ) ) );
|
|
}
|
|
|
|
void StoredTransferJob::setData( const TQByteArray& arr )
|
|
{
|
|
Q_ASSERT( m_data.isNull() ); // check that we're only called once
|
|
Q_ASSERT( m_uploadOffset == 0 ); // no upload started yet
|
|
m_data = arr;
|
|
}
|
|
|
|
void StoredTransferJob::slotStoredData( KIO::Job *, const TQByteArray &data )
|
|
{
|
|
// check for end-of-data marker:
|
|
if ( data.size() == 0 )
|
|
return;
|
|
unsigned int oldSize = m_data.size();
|
|
m_data.tqresize( oldSize + data.size(), TQGArray::SpeedOptim );
|
|
memcpy( m_data.data() + oldSize, data.data(), data.size() );
|
|
}
|
|
|
|
void StoredTransferJob::slotStoredDataReq( KIO::Job *, TQByteArray &data )
|
|
{
|
|
// Inspired from kmail's KMKernel::byteArrayToRemoteFile
|
|
// send the data in 64 KB chunks
|
|
const int MAX_CHUNK_SIZE = 64*1024;
|
|
int remainingBytes = m_data.size() - m_uploadOffset;
|
|
if( remainingBytes > MAX_CHUNK_SIZE ) {
|
|
// send MAX_CHUNK_SIZE bytes to the receiver (deep copy)
|
|
data.duplicate( m_data.data() + m_uploadOffset, MAX_CHUNK_SIZE );
|
|
m_uploadOffset += MAX_CHUNK_SIZE;
|
|
//kdDebug() << "Sending " << MAX_CHUNK_SIZE << " bytes ("
|
|
// << remainingBytes - MAX_CHUNK_SIZE << " bytes remain)\n";
|
|
} else {
|
|
// send the remaining bytes to the receiver (deep copy)
|
|
data.duplicate( m_data.data() + m_uploadOffset, remainingBytes );
|
|
m_data = TQByteArray();
|
|
m_uploadOffset = 0;
|
|
//kdDebug() << "Sending " << remainingBytes << " bytes\n";
|
|
}
|
|
}
|
|
|
|
StoredTransferJob *KIO::storedGet( const KURL& url, bool reload, bool showProgressInfo )
|
|
{
|
|
// Send decoded path and encoded query
|
|
KIO_ARGS << url;
|
|
StoredTransferJob * job = new StoredTransferJob( url, CMD_GET, packedArgs, TQByteArray(), showProgressInfo );
|
|
if (reload)
|
|
job->addMetaData("cache", "reload");
|
|
return job;
|
|
}
|
|
|
|
StoredTransferJob *KIO::storedPut( const TQByteArray& arr, const KURL& url, int permissions,
|
|
bool overwrite, bool resume, bool showProgressInfo )
|
|
{
|
|
KIO_ARGS << url << TQ_INT8( overwrite ? 1 : 0 ) << TQ_INT8( resume ? 1 : 0 ) << permissions;
|
|
StoredTransferJob * job = new StoredTransferJob( url, CMD_PUT, packedArgs, TQByteArray(), showProgressInfo );
|
|
job->setData( arr );
|
|
return job;
|
|
}
|
|
|
|
//////////
|
|
|
|
MimetypeJob::MimetypeJob( const KURL& url, int command,
|
|
const TQByteArray &packedArgs, bool showProgressInfo )
|
|
: TransferJob(url, command, packedArgs, TQByteArray(), showProgressInfo)
|
|
{
|
|
}
|
|
|
|
void MimetypeJob::start(Slave *slave)
|
|
{
|
|
TransferJob::start(slave);
|
|
}
|
|
|
|
|
|
void MimetypeJob::slotFinished( )
|
|
{
|
|
//kdDebug(7007) << "MimetypeJob::slotFinished()" << endl;
|
|
if ( m_error == KIO::ERR_IS_DIRECTORY )
|
|
{
|
|
// It is in fact a directory. This happens when HTTP redirects to FTP.
|
|
// Due to the "protocol doesn't support listing" code in KRun, we
|
|
// assumed it was a file.
|
|
kdDebug(7007) << "It is in fact a directory!" << endl;
|
|
m_mimetype = TQString::tqfromLatin1("inode/directory");
|
|
emit TransferJob::mimetype( this, m_mimetype );
|
|
m_error = 0;
|
|
}
|
|
if ( m_redirectionURL.isEmpty() || !m_redirectionURL.isValid() || m_error )
|
|
{
|
|
// Return slave to the scheduler
|
|
TransferJob::slotFinished();
|
|
} else {
|
|
//kdDebug(7007) << "MimetypeJob: Redirection to " << m_redirectionURL << endl;
|
|
if (queryMetaData("permanent-redirect")=="true")
|
|
emit permanentRedirection(this, m_url, m_redirectionURL);
|
|
staticData.truncate(0);
|
|
m_suspended = false;
|
|
m_url = m_redirectionURL;
|
|
m_redirectionURL = KURL();
|
|
m_packedArgs.truncate(0);
|
|
TQDataStream stream( m_packedArgs, IO_WriteOnly );
|
|
stream << m_url;
|
|
|
|
// Return slave to the scheduler
|
|
slaveDone();
|
|
Scheduler::doJob(this);
|
|
}
|
|
}
|
|
|
|
MimetypeJob *KIO::mimetype(const KURL& url, bool showProgressInfo )
|
|
{
|
|
KIO_ARGS << url;
|
|
MimetypeJob * job = new MimetypeJob(url, CMD_MIMETYPE, packedArgs, showProgressInfo);
|
|
if ( showProgressInfo )
|
|
Observer::self()->stating( job, url );
|
|
return job;
|
|
}
|
|
|
|
//////////////////////////
|
|
|
|
DirectCopyJob::DirectCopyJob( const KURL& url, int command,
|
|
const TQByteArray &packedArgs, bool showProgressInfo )
|
|
: SimpleJob(url, command, packedArgs, showProgressInfo)
|
|
{
|
|
}
|
|
|
|
void DirectCopyJob::start( Slave* slave )
|
|
{
|
|
connect( slave, TQT_SIGNAL(canResume( KIO::filesize_t ) ),
|
|
TQT_SLOT( slotCanResume( KIO::filesize_t ) ) );
|
|
SimpleJob::start(slave);
|
|
}
|
|
|
|
void DirectCopyJob::slotCanResume( KIO::filesize_t offset )
|
|
{
|
|
emit canResume(this, offset);
|
|
}
|
|
|
|
//////////////////////////
|
|
|
|
|
|
class FileCopyJob::FileCopyJobPrivate
|
|
{
|
|
public:
|
|
KIO::filesize_t m_sourceSize;
|
|
time_t m_modificationTime;
|
|
SimpleJob *m_delJob;
|
|
};
|
|
|
|
/*
|
|
* The FileCopyJob works according to the famous Bayern
|
|
* 'Alternating Bitburger Protocol': we either drink a beer or we
|
|
* we order a beer, but never both at the same time.
|
|
* Tranlated to io-slaves: We alternate between receiving a block of data
|
|
* and sending it away.
|
|
*/
|
|
FileCopyJob::FileCopyJob( const KURL& src, const KURL& dest, int permissions,
|
|
bool move, bool overwrite, bool resume, bool showProgressInfo)
|
|
: Job(showProgressInfo), m_src(src), m_dest(dest),
|
|
m_permissions(permissions), m_move(move), m_overwrite(overwrite), m_resume(resume),
|
|
m_totalSize(0)
|
|
{
|
|
if (showProgressInfo && !move)
|
|
Observer::self()->slotCopying( this, src, dest );
|
|
else if (showProgressInfo && move)
|
|
Observer::self()->slotMoving( this, src, dest );
|
|
|
|
//kdDebug(7007) << "FileCopyJob::FileCopyJob()" << endl;
|
|
m_moveJob = 0;
|
|
m_copyJob = 0;
|
|
m_getJob = 0;
|
|
m_putJob = 0;
|
|
d = new FileCopyJobPrivate;
|
|
d->m_delJob = 0;
|
|
d->m_sourceSize = (KIO::filesize_t) -1;
|
|
d->m_modificationTime = static_cast<time_t>( -1 );
|
|
TQTimer::singleShot(0, this, TQT_SLOT(slotStart()));
|
|
}
|
|
|
|
void FileCopyJob::slotStart()
|
|
{
|
|
if ( m_move )
|
|
{
|
|
// The if() below must be the same as the one in startBestCopyMethod
|
|
if ((m_src.protocol() == m_dest.protocol()) &&
|
|
(m_src.host() == m_dest.host()) &&
|
|
(m_src.port() == m_dest.port()) &&
|
|
(m_src.user() == m_dest.user()) &&
|
|
(m_src.pass() == m_dest.pass()) &&
|
|
!m_src.hasSubURL() && !m_dest.hasSubURL())
|
|
{
|
|
startRenameJob(m_src);
|
|
return;
|
|
}
|
|
else if (m_src.isLocalFile() && KProtocolInfo::canRenameFromFile(m_dest))
|
|
{
|
|
startRenameJob(m_dest);
|
|
return;
|
|
}
|
|
else if (m_dest.isLocalFile() && KProtocolInfo::canRenameToFile(m_src))
|
|
{
|
|
startRenameJob(m_src);
|
|
return;
|
|
}
|
|
// No fast-move available, use copy + del.
|
|
}
|
|
startBestCopyMethod();
|
|
}
|
|
|
|
void FileCopyJob::startBestCopyMethod()
|
|
{
|
|
if ((m_src.protocol() == m_dest.protocol()) &&
|
|
(m_src.host() == m_dest.host()) &&
|
|
(m_src.port() == m_dest.port()) &&
|
|
(m_src.user() == m_dest.user()) &&
|
|
(m_src.pass() == m_dest.pass()) &&
|
|
!m_src.hasSubURL() && !m_dest.hasSubURL())
|
|
{
|
|
startCopyJob();
|
|
}
|
|
else if (m_src.isLocalFile() && KProtocolInfo::canCopyFromFile(m_dest))
|
|
{
|
|
startCopyJob(m_dest);
|
|
}
|
|
else if (m_dest.isLocalFile() && KProtocolInfo::canCopyToFile(m_src))
|
|
{
|
|
startCopyJob(m_src);
|
|
}
|
|
else
|
|
{
|
|
startDataPump();
|
|
}
|
|
}
|
|
|
|
FileCopyJob::~FileCopyJob()
|
|
{
|
|
delete d;
|
|
}
|
|
|
|
void FileCopyJob::setSourceSize( off_t size )
|
|
{
|
|
d->m_sourceSize = size;
|
|
if (size != (off_t) -1)
|
|
m_totalSize = size;
|
|
}
|
|
|
|
void FileCopyJob::setSourceSize64( KIO::filesize_t size )
|
|
{
|
|
d->m_sourceSize = size;
|
|
if (size != (KIO::filesize_t) -1)
|
|
m_totalSize = size;
|
|
}
|
|
|
|
void FileCopyJob::setModificationTime( time_t mtime )
|
|
{
|
|
d->m_modificationTime = mtime;
|
|
}
|
|
|
|
void FileCopyJob::startCopyJob()
|
|
{
|
|
startCopyJob(m_src);
|
|
}
|
|
|
|
void FileCopyJob::startCopyJob(const KURL &slave_url)
|
|
{
|
|
//kdDebug(7007) << "FileCopyJob::startCopyJob()" << endl;
|
|
KIO_ARGS << m_src << m_dest << m_permissions << (TQ_INT8) m_overwrite;
|
|
m_copyJob = new DirectCopyJob(slave_url, CMD_COPY, packedArgs, false);
|
|
addSubjob( m_copyJob );
|
|
connectSubjob( m_copyJob );
|
|
connect( m_copyJob, TQT_SIGNAL(canResume(KIO::Job *, KIO::filesize_t)),
|
|
TQT_SLOT( slotCanResume(KIO::Job *, KIO::filesize_t)));
|
|
}
|
|
|
|
void FileCopyJob::startRenameJob(const KURL &slave_url)
|
|
{
|
|
KIO_ARGS << m_src << m_dest << (TQ_INT8) m_overwrite;
|
|
m_moveJob = new SimpleJob(slave_url, CMD_RENAME, packedArgs, false);
|
|
addSubjob( m_moveJob );
|
|
connectSubjob( m_moveJob );
|
|
}
|
|
|
|
void FileCopyJob::connectSubjob( SimpleJob * job )
|
|
{
|
|
connect( job, TQT_SIGNAL(totalSize( KIO::Job*, KIO::filesize_t )),
|
|
this, TQT_SLOT( slotTotalSize(KIO::Job*, KIO::filesize_t)) );
|
|
|
|
connect( job, TQT_SIGNAL(processedSize( KIO::Job*, KIO::filesize_t )),
|
|
this, TQT_SLOT( slotProcessedSize(KIO::Job*, KIO::filesize_t)) );
|
|
|
|
connect( job, TQT_SIGNAL(percent( KIO::Job*, unsigned long )),
|
|
this, TQT_SLOT( slotPercent(KIO::Job*, unsigned long)) );
|
|
|
|
}
|
|
|
|
void FileCopyJob::slotProcessedSize( KIO::Job *, KIO::filesize_t size )
|
|
{
|
|
setProcessedSize(size);
|
|
emit processedSize( this, size );
|
|
if ( size > m_totalSize ) {
|
|
slotTotalSize( this, size ); // safety
|
|
}
|
|
emitPercent( size, m_totalSize );
|
|
}
|
|
|
|
void FileCopyJob::slotTotalSize( KIO::Job*, KIO::filesize_t size )
|
|
{
|
|
if (size > m_totalSize)
|
|
{
|
|
m_totalSize = size;
|
|
emit totalSize( this, m_totalSize );
|
|
}
|
|
}
|
|
|
|
void FileCopyJob::slotPercent( KIO::Job*, unsigned long pct )
|
|
{
|
|
if ( pct > m_percent )
|
|
{
|
|
m_percent = pct;
|
|
emit percent( this, m_percent );
|
|
}
|
|
}
|
|
|
|
void FileCopyJob::startDataPump()
|
|
{
|
|
//kdDebug(7007) << "FileCopyJob::startDataPump()" << endl;
|
|
|
|
m_canResume = false;
|
|
m_resumeAnswerSent = false;
|
|
m_getJob = 0L; // for now
|
|
m_putJob = put( m_dest, m_permissions, m_overwrite, m_resume, false /* no GUI */);
|
|
if ( d->m_modificationTime != static_cast<time_t>( -1 ) ) {
|
|
TQDateTime dt; dt.setTime_t( d->m_modificationTime );
|
|
m_putJob->addMetaData( "modified", dt.toString( Qt::ISODate ) );
|
|
}
|
|
//kdDebug(7007) << "FileCopyJob: m_putJob = " << m_putJob << " m_dest=" << m_dest << endl;
|
|
|
|
// The first thing the put job will tell us is whether we can
|
|
// resume or not (this is always emitted)
|
|
connect( m_putJob, TQT_SIGNAL(canResume(KIO::Job *, KIO::filesize_t)),
|
|
TQT_SLOT( slotCanResume(KIO::Job *, KIO::filesize_t)));
|
|
connect( m_putJob, TQT_SIGNAL(dataReq(KIO::Job *, TQByteArray&)),
|
|
TQT_SLOT( slotDataReq(KIO::Job *, TQByteArray&)));
|
|
addSubjob( m_putJob );
|
|
}
|
|
|
|
void FileCopyJob::slotCanResume( KIO::Job* job, KIO::filesize_t offset )
|
|
{
|
|
if ( job == m_putJob || job == m_copyJob )
|
|
{
|
|
//kdDebug(7007) << "FileCopyJob::slotCanResume from PUT job. offset=" << KIO::number(offset) << endl;
|
|
if (offset)
|
|
{
|
|
RenameDlg_Result res = R_RESUME;
|
|
|
|
if (!KProtocolManager::autoResume() && !m_overwrite)
|
|
{
|
|
TQString newPath;
|
|
KIO::Job* job = ( !m_progressId && parentJob() ) ? parentJob() : this;
|
|
// Ask confirmation about resuming previous transfer
|
|
res = Observer::self()->open_RenameDlg(
|
|
job, i18n("File Already Exists"),
|
|
m_src.url(),
|
|
m_dest.url(),
|
|
(RenameDlg_Mode) (M_OVERWRITE | M_RESUME | M_NORENAME), newPath,
|
|
d->m_sourceSize, offset );
|
|
}
|
|
|
|
if ( res == R_OVERWRITE || m_overwrite )
|
|
offset = 0;
|
|
else if ( res == R_CANCEL )
|
|
{
|
|
if ( job == m_putJob )
|
|
m_putJob->kill(true);
|
|
else
|
|
m_copyJob->kill(true);
|
|
m_error = ERR_USER_CANCELED;
|
|
emitResult();
|
|
return;
|
|
}
|
|
}
|
|
else
|
|
m_resumeAnswerSent = true; // No need for an answer
|
|
|
|
if ( job == m_putJob )
|
|
{
|
|
m_getJob = get( m_src, false, false /* no GUI */ );
|
|
//kdDebug(7007) << "FileCopyJob: m_getJob = " << m_getJob << endl;
|
|
m_getJob->addMetaData( "errorPage", "false" );
|
|
m_getJob->addMetaData( "AllowCompressedPage", "false" );
|
|
// Set size in subjob. This helps if the slave doesn't emit totalSize.
|
|
if ( d->m_sourceSize != (KIO::filesize_t)-1 )
|
|
m_getJob->slotTotalSize( d->m_sourceSize );
|
|
if (offset)
|
|
{
|
|
//kdDebug(7007) << "Setting metadata for resume to " << (unsigned long) offset << endl;
|
|
// TODO KDE4: rename to seek or offset and document it
|
|
// This isn't used only for resuming, but potentially also for extracting (#72302).
|
|
m_getJob->addMetaData( "resume", KIO::number(offset) );
|
|
|
|
// Might or might not get emitted
|
|
connect( m_getJob, TQT_SIGNAL(canResume(KIO::Job *, KIO::filesize_t)),
|
|
TQT_SLOT( slotCanResume(KIO::Job *, KIO::filesize_t)));
|
|
}
|
|
m_putJob->slave()->setOffset( offset );
|
|
|
|
m_putJob->suspend();
|
|
addSubjob( m_getJob );
|
|
connectSubjob( m_getJob ); // Progress info depends on get
|
|
m_getJob->resume(); // Order a beer
|
|
|
|
connect( m_getJob, TQT_SIGNAL(data(KIO::Job*,const TQByteArray&)),
|
|
TQT_SLOT( slotData(KIO::Job*,const TQByteArray&)) );
|
|
connect( m_getJob, TQT_SIGNAL(mimetype(KIO::Job*,const TQString&) ),
|
|
TQT_SLOT(slotMimetype(KIO::Job*,const TQString&)) );
|
|
}
|
|
else // copyjob
|
|
{
|
|
m_copyJob->slave()->sendResumeAnswer( offset != 0 );
|
|
}
|
|
}
|
|
else if ( job == m_getJob )
|
|
{
|
|
// Cool, the get job said ok, we can resume
|
|
m_canResume = true;
|
|
//kdDebug(7007) << "FileCopyJob::slotCanResume from the GET job -> we can resume" << endl;
|
|
|
|
m_getJob->slave()->setOffset( m_putJob->slave()->offset() );
|
|
}
|
|
else
|
|
kdWarning(7007) << "FileCopyJob::slotCanResume from unknown job=" << job
|
|
<< " m_getJob=" << m_getJob << " m_putJob=" << m_putJob << endl;
|
|
}
|
|
|
|
void FileCopyJob::slotData( KIO::Job * , const TQByteArray &data)
|
|
{
|
|
//kdDebug(7007) << "FileCopyJob::slotData" << endl;
|
|
//kdDebug(7007) << " data size : " << data.size() << endl;
|
|
assert(m_putJob);
|
|
if (!m_putJob) return; // Don't crash
|
|
m_getJob->suspend();
|
|
m_putJob->resume(); // Drink the beer
|
|
m_buffer = data;
|
|
|
|
// On the first set of data incoming, we tell the "put" slave about our
|
|
// decision about resuming
|
|
if (!m_resumeAnswerSent)
|
|
{
|
|
m_resumeAnswerSent = true;
|
|
//kdDebug(7007) << "FileCopyJob::slotData (first time) -> send resume answer " << m_canResume << endl;
|
|
m_putJob->slave()->sendResumeAnswer( m_canResume );
|
|
}
|
|
}
|
|
|
|
void FileCopyJob::slotDataReq( KIO::Job * , TQByteArray &data)
|
|
{
|
|
//kdDebug(7007) << "FileCopyJob::slotDataReq" << endl;
|
|
if (!m_resumeAnswerSent && !m_getJob)
|
|
{
|
|
// This can't happen (except as a migration bug on 12/10/2000)
|
|
m_error = ERR_INTERNAL;
|
|
m_errorText = "'Put' job didn't send canResume or 'Get' job didn't send data!";
|
|
m_putJob->kill(true);
|
|
emitResult();
|
|
return;
|
|
}
|
|
if (m_getJob)
|
|
{
|
|
m_getJob->resume(); // Order more beer
|
|
m_putJob->suspend();
|
|
}
|
|
data = m_buffer;
|
|
m_buffer = TQByteArray();
|
|
}
|
|
|
|
void FileCopyJob::slotMimetype( KIO::Job*, const TQString& type )
|
|
{
|
|
emit mimetype( this, type );
|
|
}
|
|
|
|
void FileCopyJob::slotResult( KIO::Job *job)
|
|
{
|
|
//kdDebug(7007) << "FileCopyJob this=" << this << " ::slotResult(" << job << ")" << endl;
|
|
// Did job have an error ?
|
|
if ( job->error() )
|
|
{
|
|
if ((job == m_moveJob) && (job->error() == ERR_UNSUPPORTED_ACTION))
|
|
{
|
|
m_moveJob = 0;
|
|
startBestCopyMethod();
|
|
removeSubjob(job);
|
|
return;
|
|
}
|
|
else if ((job == m_copyJob) && (job->error() == ERR_UNSUPPORTED_ACTION))
|
|
{
|
|
m_copyJob = 0;
|
|
startDataPump();
|
|
removeSubjob(job);
|
|
return;
|
|
}
|
|
else if (job == m_getJob)
|
|
{
|
|
m_getJob = 0L;
|
|
if (m_putJob)
|
|
m_putJob->kill(true);
|
|
}
|
|
else if (job == m_putJob)
|
|
{
|
|
m_putJob = 0L;
|
|
if (m_getJob)
|
|
m_getJob->kill(true);
|
|
}
|
|
m_error = job->error();
|
|
m_errorText = job->errorText();
|
|
emitResult();
|
|
return;
|
|
}
|
|
|
|
if (job == m_moveJob)
|
|
{
|
|
m_moveJob = 0; // Finished
|
|
}
|
|
|
|
if (job == m_copyJob)
|
|
{
|
|
m_copyJob = 0;
|
|
if (m_move)
|
|
{
|
|
d->m_delJob = file_delete( m_src, false/*no GUI*/ ); // Delete source
|
|
addSubjob(d->m_delJob);
|
|
}
|
|
}
|
|
|
|
if (job == m_getJob)
|
|
{
|
|
m_getJob = 0; // No action required
|
|
if (m_putJob)
|
|
m_putJob->resume();
|
|
}
|
|
|
|
if (job == m_putJob)
|
|
{
|
|
//kdDebug(7007) << "FileCopyJob: m_putJob finished " << endl;
|
|
m_putJob = 0;
|
|
if (m_getJob)
|
|
{
|
|
kdWarning(7007) << "WARNING ! Get still going on..." << endl;
|
|
m_getJob->resume();
|
|
}
|
|
if (m_move)
|
|
{
|
|
d->m_delJob = file_delete( m_src, false/*no GUI*/ ); // Delete source
|
|
addSubjob(d->m_delJob);
|
|
}
|
|
}
|
|
|
|
if (job == d->m_delJob)
|
|
{
|
|
d->m_delJob = 0; // Finished
|
|
}
|
|
removeSubjob(job);
|
|
}
|
|
|
|
FileCopyJob *KIO::file_copy( const KURL& src, const KURL& dest, int permissions,
|
|
bool overwrite, bool resume, bool showProgressInfo)
|
|
{
|
|
return new FileCopyJob( src, dest, permissions, false, overwrite, resume, showProgressInfo );
|
|
}
|
|
|
|
FileCopyJob *KIO::file_move( const KURL& src, const KURL& dest, int permissions,
|
|
bool overwrite, bool resume, bool showProgressInfo)
|
|
{
|
|
return new FileCopyJob( src, dest, permissions, true, overwrite, resume, showProgressInfo );
|
|
}
|
|
|
|
SimpleJob *KIO::file_delete( const KURL& src, bool showProgressInfo)
|
|
{
|
|
KIO_ARGS << src << TQ_INT8(true); // isFile
|
|
return new SimpleJob(src, CMD_DEL, packedArgs, showProgressInfo );
|
|
}
|
|
|
|
//////////
|
|
|
|
// KDE 4: Make it const TQString & _prefix
|
|
ListJob::ListJob(const KURL& u, bool showProgressInfo, bool _recursive, TQString _prefix, bool _includeHidden) :
|
|
SimpleJob(u, CMD_LISTDIR, TQByteArray(), showProgressInfo),
|
|
recursive(_recursive), includeHidden(_includeHidden), prefix(_prefix), m_processedEntries(0)
|
|
{
|
|
// We couldn't set the args when calling the parent constructor,
|
|
// so do it now.
|
|
TQDataStream stream( m_packedArgs, IO_WriteOnly );
|
|
stream << u;
|
|
}
|
|
|
|
void ListJob::slotListEntries( const KIO::UDSEntryList& list )
|
|
{
|
|
// Emit progress info (takes care of emit processedSize and percent)
|
|
m_processedEntries += list.count();
|
|
slotProcessedSize( m_processedEntries );
|
|
|
|
if (recursive) {
|
|
UDSEntryListConstIterator it = list.begin();
|
|
UDSEntryListConstIterator end = list.end();
|
|
|
|
for (; it != end; ++it) {
|
|
bool isDir = false;
|
|
bool isLink = false;
|
|
KURL itemURL;
|
|
|
|
UDSEntry::ConstIterator it2 = (*it).begin();
|
|
UDSEntry::ConstIterator end2 = (*it).end();
|
|
for( ; it2 != end2; it2++ ) {
|
|
switch( (*it2).m_uds ) {
|
|
case UDS_FILE_TYPE:
|
|
isDir = S_ISDIR((*it2).m_long);
|
|
break;
|
|
case UDS_NAME:
|
|
if( itemURL.isEmpty() ) {
|
|
itemURL = url();
|
|
itemURL.addPath( (*it2).m_str );
|
|
}
|
|
break;
|
|
case UDS_URL:
|
|
itemURL = (*it2).m_str;
|
|
break;
|
|
case UDS_LINK_DEST:
|
|
// This is a link !!! Don't follow !
|
|
isLink = !(*it2).m_str.isEmpty();
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
if (isDir && !isLink) {
|
|
const TQString filename = itemURL.fileName();
|
|
// skip hidden dirs when listing if requested
|
|
if (filename != ".." && filename != "." && (includeHidden || filename[0] != '.')) {
|
|
ListJob *job = new ListJob(itemURL,
|
|
false /*no progress info!*/,
|
|
true /*recursive*/,
|
|
prefix + filename + "/",
|
|
includeHidden);
|
|
Scheduler::scheduleJob(job);
|
|
connect(job, TQT_SIGNAL(entries( KIO::Job *,
|
|
const KIO::UDSEntryList& )),
|
|
TQT_SLOT( gotEntries( KIO::Job*,
|
|
const KIO::UDSEntryList& )));
|
|
addSubjob(job);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Not recursive, or top-level of recursive listing : return now (send . and .. as well)
|
|
// exclusion of hidden files also requires the full sweep, but the case for full-listing
|
|
// a single dir is probably common enough to justify the shortcut
|
|
if (prefix.isNull() && includeHidden) {
|
|
emit entries(this, list);
|
|
} else {
|
|
// cull the unwanted hidden dirs and/or parent dir references from the listing, then emit that
|
|
UDSEntryList newlist;
|
|
|
|
UDSEntryListConstIterator it = list.begin();
|
|
UDSEntryListConstIterator end = list.end();
|
|
for (; it != end; ++it) {
|
|
|
|
UDSEntry newone = *it;
|
|
UDSEntry::Iterator it2 = newone.begin();
|
|
TQString filename;
|
|
for( ; it2 != newone.end(); it2++ ) {
|
|
if ((*it2).m_uds == UDS_NAME) {
|
|
filename = (*it2).m_str;
|
|
(*it2).m_str = prefix + filename;
|
|
}
|
|
}
|
|
// Avoid returning entries like subdir/. and subdir/.., but include . and .. for
|
|
// the toplevel dir, and skip hidden files/dirs if that was requested
|
|
if ( (prefix.isNull() || (filename != ".." && filename != ".") )
|
|
&& (includeHidden || (filename[0] != '.') ) )
|
|
newlist.append(newone);
|
|
}
|
|
|
|
emit entries(this, newlist);
|
|
}
|
|
}
|
|
|
|
void ListJob::gotEntries(KIO::Job *, const KIO::UDSEntryList& list )
|
|
{
|
|
// Forward entries received by subjob - faking we received them ourselves
|
|
emit entries(this, list);
|
|
}
|
|
|
|
void ListJob::slotResult( KIO::Job * job )
|
|
{
|
|
// If we can't list a subdir, the result is still ok
|
|
// This is why we override Job::slotResult() - to skip error checking
|
|
removeSubjob( job );
|
|
}
|
|
|
|
void ListJob::slotRedirection( const KURL & url )
|
|
{
|
|
if (!kapp->authorizeURLAction("redirect", m_url, url))
|
|
{
|
|
kdWarning(7007) << "ListJob: Redirection from " << m_url << " to " << url << " REJECTED!" << endl;
|
|
return;
|
|
}
|
|
m_redirectionURL = url; // We'll remember that when the job finishes
|
|
if (m_url.hasUser() && !url.hasUser() && (m_url.host().lower() == url.host().lower()))
|
|
m_redirectionURL.setUser(m_url.user()); // Preserve user
|
|
emit redirection( this, m_redirectionURL );
|
|
}
|
|
|
|
void ListJob::slotFinished()
|
|
{
|
|
// Support for listing archives as directories
|
|
if ( m_error == KIO::ERR_IS_FILE && m_url.isLocalFile() ) {
|
|
KMimeType::Ptr ptr = KMimeType::findByURL( m_url, 0, true, true );
|
|
if ( ptr ) {
|
|
TQString proto = ptr->property("X-KDE-LocalProtocol").toString();
|
|
if ( !proto.isEmpty() && KProtocolInfo::isKnownProtocol(proto) ) {
|
|
m_redirectionURL = m_url;
|
|
m_redirectionURL.setProtocol( proto );
|
|
m_error = 0;
|
|
emit redirection(this,m_redirectionURL);
|
|
}
|
|
}
|
|
}
|
|
if ( m_redirectionURL.isEmpty() || !m_redirectionURL.isValid() || m_error ) {
|
|
// Return slave to the scheduler
|
|
SimpleJob::slotFinished();
|
|
} else {
|
|
|
|
//kdDebug(7007) << "ListJob: Redirection to " << m_redirectionURL << endl;
|
|
if (queryMetaData("permanent-redirect")=="true")
|
|
emit permanentRedirection(this, m_url, m_redirectionURL);
|
|
m_url = m_redirectionURL;
|
|
m_redirectionURL = KURL();
|
|
m_packedArgs.truncate(0);
|
|
TQDataStream stream( m_packedArgs, IO_WriteOnly );
|
|
stream << m_url;
|
|
|
|
// Return slave to the scheduler
|
|
slaveDone();
|
|
Scheduler::doJob(this);
|
|
}
|
|
}
|
|
|
|
void ListJob::slotMetaData( const KIO::MetaData &_metaData) {
|
|
SimpleJob::slotMetaData(_metaData);
|
|
storeSSLSessionFromJob(m_redirectionURL);
|
|
}
|
|
|
|
ListJob *KIO::listDir( const KURL& url, bool showProgressInfo, bool includeHidden )
|
|
{
|
|
ListJob * job = new ListJob(url, showProgressInfo,false,TQString::null,includeHidden);
|
|
return job;
|
|
}
|
|
|
|
ListJob *KIO::listRecursive( const KURL& url, bool showProgressInfo, bool includeHidden )
|
|
{
|
|
ListJob * job = new ListJob(url, showProgressInfo, true,TQString::null,includeHidden);
|
|
return job;
|
|
}
|
|
|
|
void ListJob::setUnrestricted(bool unrestricted)
|
|
{
|
|
if (unrestricted)
|
|
extraFlags() |= EF_ListJobUnrestricted;
|
|
else
|
|
extraFlags() &= ~EF_ListJobUnrestricted;
|
|
}
|
|
|
|
void ListJob::start(Slave *slave)
|
|
{
|
|
if (kapp && !kapp->authorizeURLAction("list", m_url, m_url) && !(extraFlags() & EF_ListJobUnrestricted))
|
|
{
|
|
m_error = ERR_ACCESS_DENIED;
|
|
m_errorText = m_url.url();
|
|
TQTimer::singleShot(0, this, TQT_SLOT(slotFinished()) );
|
|
return;
|
|
}
|
|
connect( slave, TQT_SIGNAL( listEntries( const KIO::UDSEntryList& )),
|
|
TQT_SLOT( slotListEntries( const KIO::UDSEntryList& )));
|
|
connect( slave, TQT_SIGNAL( totalSize( KIO::filesize_t ) ),
|
|
TQT_SLOT( slotTotalSize( KIO::filesize_t ) ) );
|
|
connect( slave, TQT_SIGNAL( redirection(const KURL &) ),
|
|
TQT_SLOT( slotRedirection(const KURL &) ) );
|
|
|
|
SimpleJob::start(slave);
|
|
}
|
|
|
|
class CopyJob::CopyJobPrivate
|
|
{
|
|
public:
|
|
CopyJobPrivate() {
|
|
m_defaultPermissions = false;
|
|
m_bURLDirty = false;
|
|
}
|
|
// This is the dest URL that was initially given to CopyJob
|
|
// It is copied into m_dest, which can be changed for a given src URL
|
|
// (when using the RENAME dialog in slotResult),
|
|
// and which will be reset for the next src URL.
|
|
KURL m_globalDest;
|
|
// The state info about that global dest
|
|
CopyJob::DestinationState m_globalDestinationState;
|
|
// See setDefaultPermissions
|
|
bool m_defaultPermissions;
|
|
// Whether URLs changed (and need to be emitted by the next slotReport call)
|
|
bool m_bURLDirty;
|
|
// Used after copying all the files into the dirs, to set mtime (TODO: and permissions?)
|
|
// after the copy is done
|
|
TQValueList<CopyInfo> m_directoriesCopied;
|
|
};
|
|
|
|
CopyJob::CopyJob( const KURL::List& src, const KURL& dest, CopyMode mode, bool asMethod, bool showProgressInfo )
|
|
: Job(showProgressInfo), m_mode(mode), m_asMethod(asMethod),
|
|
destinationState(DEST_NOT_STATED), state(STATE_STATING),
|
|
m_totalSize(0), m_processedSize(0), m_fileProcessedSize(0),
|
|
m_processedFiles(0), m_processedDirs(0),
|
|
m_srcList(src), m_currentStatSrc(m_srcList.begin()),
|
|
m_bCurrentOperationIsLink(false), m_bSingleFileCopy(false), m_bOnlyRenames(mode==Move),
|
|
m_dest(dest), m_bAutoSkip( false ), m_bOverwriteAll( false ),
|
|
m_conflictError(0), m_reportTimer(0)
|
|
{
|
|
d = new CopyJobPrivate;
|
|
d->m_globalDest = dest;
|
|
d->m_globalDestinationState = destinationState;
|
|
|
|
if ( showProgressInfo ) {
|
|
connect( this, TQT_SIGNAL( totalFiles( KIO::Job*, unsigned long ) ),
|
|
Observer::self(), TQT_SLOT( slotTotalFiles( KIO::Job*, unsigned long ) ) );
|
|
|
|
connect( this, TQT_SIGNAL( totalDirs( KIO::Job*, unsigned long ) ),
|
|
Observer::self(), TQT_SLOT( slotTotalDirs( KIO::Job*, unsigned long ) ) );
|
|
}
|
|
TQTimer::singleShot(0, this, TQT_SLOT(slotStart()));
|
|
/**
|
|
States:
|
|
STATE_STATING for the dest
|
|
STATE_STATING for each src url (statNextSrc)
|
|
for each: if dir -> STATE_LISTING (filling 'dirs' and 'files')
|
|
but if direct rename possible: STATE_RENAMING instead.
|
|
STATE_CREATING_DIRS (createNextDir, iterating over 'dirs')
|
|
if conflict: STATE_CONFLICT_CREATING_DIRS
|
|
STATE_COPYING_FILES (copyNextFile, iterating over 'files')
|
|
if conflict: STATE_CONFLICT_COPYING_FILES
|
|
STATE_DELETING_DIRS (deleteNextDir) (if moving)
|
|
STATE_SETTING_DIR_ATTRIBUTES (setNextDirAttribute, iterating over d->m_directoriesCopied)
|
|
done.
|
|
*/
|
|
}
|
|
|
|
CopyJob::~CopyJob()
|
|
{
|
|
delete d;
|
|
}
|
|
|
|
void CopyJob::slotStart()
|
|
{
|
|
/**
|
|
We call the functions directly instead of using signals.
|
|
Calling a function via a signal takes approx. 65 times the time
|
|
compared to calling it directly (at least on my machine). aleXXX
|
|
*/
|
|
m_reportTimer = new TQTimer(this);
|
|
|
|
connect(m_reportTimer,TQT_SIGNAL(timeout()),this,TQT_SLOT(slotReport()));
|
|
m_reportTimer->start(REPORT_TIMEOUT,false);
|
|
|
|
// Stat the dest
|
|
KIO::Job * job = KIO::stat( m_dest, false, 2, false );
|
|
//kdDebug(7007) << "CopyJob:stating the dest " << m_dest << endl;
|
|
addSubjob(job);
|
|
}
|
|
|
|
// For unit test purposes
|
|
KIO_EXPORT bool kio_resolve_local_urls = true;
|
|
|
|
void CopyJob::slotResultStating( Job *job )
|
|
{
|
|
//kdDebug(7007) << "CopyJob::slotResultStating" << endl;
|
|
// Was there an error while stating the src ?
|
|
if (job->error() && destinationState != DEST_NOT_STATED )
|
|
{
|
|
KURL srcurl = ((SimpleJob*)job)->url();
|
|
if ( !srcurl.isLocalFile() )
|
|
{
|
|
// Probably : src doesn't exist. Well, over some protocols (e.g. FTP)
|
|
// this info isn't really reliable (thanks to MS FTP servers).
|
|
// We'll assume a file, and try to download anyway.
|
|
kdDebug(7007) << "Error while stating source. Activating hack" << endl;
|
|
subjobs.remove( job );
|
|
assert ( subjobs.isEmpty() ); // We should have only one job at a time ...
|
|
struct CopyInfo info;
|
|
info.permissions = (mode_t) -1;
|
|
info.mtime = (time_t) -1;
|
|
info.ctime = (time_t) -1;
|
|
info.size = (KIO::filesize_t)-1;
|
|
info.uSource = srcurl;
|
|
info.uDest = m_dest;
|
|
// Append filename or dirname to destination URL, if allowed
|
|
if ( destinationState == DEST_IS_DIR && !m_asMethod )
|
|
info.uDest.addPath( srcurl.fileName() );
|
|
|
|
files.append( info );
|
|
statNextSrc();
|
|
return;
|
|
}
|
|
// Local file. If stat fails, the file definitely doesn't exist.
|
|
Job::slotResult( job ); // will set the error and emit result(this)
|
|
return;
|
|
}
|
|
|
|
// Is it a file or a dir ? Does it have a local path?
|
|
UDSEntry entry = ((StatJob*)job)->statResult();
|
|
bool bDir = false;
|
|
bool bLink = false;
|
|
TQString sName;
|
|
TQString sLocalPath;
|
|
UDSEntry::ConstIterator it2 = entry.begin();
|
|
for( ; it2 != entry.end(); it2++ ) {
|
|
if ( ((*it2).m_uds) == UDS_FILE_TYPE )
|
|
bDir = S_ISDIR( (mode_t)(*it2).m_long );
|
|
else if ( ((*it2).m_uds) == UDS_LINK_DEST )
|
|
bLink = !((*it2).m_str.isEmpty());
|
|
else if ( ((*it2).m_uds) == UDS_NAME )
|
|
sName = (*it2).m_str;
|
|
else if ( ((*it2).m_uds) == UDS_LOCAL_PATH )
|
|
sLocalPath = (*it2).m_str;
|
|
}
|
|
|
|
if ( destinationState == DEST_NOT_STATED )
|
|
// we were stating the dest
|
|
{
|
|
if (job->error())
|
|
destinationState = DEST_DOESNT_EXIST;
|
|
else {
|
|
// Treat symlinks to dirs as dirs here, so no test on bLink
|
|
destinationState = bDir ? DEST_IS_DIR : DEST_IS_FILE;
|
|
//kdDebug(7007) << "CopyJob::slotResultStating dest is dir:" << bDir << endl;
|
|
}
|
|
const bool isGlobalDest = m_dest == d->m_globalDest;
|
|
if ( isGlobalDest )
|
|
d->m_globalDestinationState = destinationState;
|
|
|
|
if ( !sLocalPath.isEmpty() && kio_resolve_local_urls ) {
|
|
m_dest = KURL();
|
|
m_dest.setPath(sLocalPath);
|
|
if ( isGlobalDest )
|
|
d->m_globalDest = m_dest;
|
|
}
|
|
|
|
subjobs.remove( job );
|
|
assert ( subjobs.isEmpty() );
|
|
|
|
// After knowing what the dest is, we can start stat'ing the first src.
|
|
statCurrentSrc();
|
|
return;
|
|
}
|
|
// We were stating the current source URL
|
|
m_currentDest = m_dest; // used by slotEntries
|
|
// Create a dummy list with it, for slotEntries
|
|
UDSEntryList lst;
|
|
lst.append(entry);
|
|
|
|
// There 6 cases, and all end up calling slotEntries(job, lst) first :
|
|
// 1 - src is a dir, destination is a directory,
|
|
// slotEntries will append the source-dir-name to the destination
|
|
// 2 - src is a dir, destination is a file, ERROR (done later on)
|
|
// 3 - src is a dir, destination doesn't exist, then it's the destination dirname,
|
|
// so slotEntries will use it as destination.
|
|
|
|
// 4 - src is a file, destination is a directory,
|
|
// slotEntries will append the filename to the destination.
|
|
// 5 - src is a file, destination is a file, m_dest is the exact destination name
|
|
// 6 - src is a file, destination doesn't exist, m_dest is the exact destination name
|
|
// Tell slotEntries not to alter the src url
|
|
m_bCurrentSrcIsDir = false;
|
|
slotEntries(job, lst);
|
|
|
|
KURL srcurl;
|
|
if (!sLocalPath.isEmpty())
|
|
srcurl.setPath(sLocalPath);
|
|
else
|
|
srcurl = ((SimpleJob*)job)->url();
|
|
|
|
subjobs.remove( job );
|
|
assert ( subjobs.isEmpty() ); // We should have only one job at a time ...
|
|
|
|
if ( bDir
|
|
&& !bLink // treat symlinks as files (no recursion)
|
|
&& m_mode != Link ) // No recursion in Link mode either.
|
|
{
|
|
//kdDebug(7007) << " Source is a directory " << endl;
|
|
|
|
m_bCurrentSrcIsDir = true; // used by slotEntries
|
|
if ( destinationState == DEST_IS_DIR ) // (case 1)
|
|
{
|
|
if ( !m_asMethod )
|
|
{
|
|
// Use <desturl>/<directory_copied> as destination, from now on
|
|
TQString directory = srcurl.fileName();
|
|
if ( !sName.isEmpty() && KProtocolInfo::fileNameUsedForCopying( srcurl ) == KProtocolInfo::Name )
|
|
{
|
|
directory = sName;
|
|
}
|
|
m_currentDest.addPath( directory );
|
|
}
|
|
}
|
|
else if ( destinationState == DEST_IS_FILE ) // (case 2)
|
|
{
|
|
m_error = ERR_IS_FILE;
|
|
m_errorText = m_dest.prettyURL();
|
|
emitResult();
|
|
return;
|
|
}
|
|
else // (case 3)
|
|
{
|
|
// otherwise dest is new name for toplevel dir
|
|
// so the destination exists, in fact, from now on.
|
|
// (This even works with other src urls in the list, since the
|
|
// dir has effectively been created)
|
|
destinationState = DEST_IS_DIR;
|
|
if ( m_dest == d->m_globalDest )
|
|
d->m_globalDestinationState = destinationState;
|
|
}
|
|
|
|
startListing( srcurl );
|
|
}
|
|
else
|
|
{
|
|
//kdDebug(7007) << " Source is a file (or a symlink), or we are linking -> no recursive listing " << endl;
|
|
statNextSrc();
|
|
}
|
|
}
|
|
|
|
void CopyJob::slotReport()
|
|
{
|
|
// If showProgressInfo was set, m_progressId is > 0.
|
|
Observer * observer = m_progressId ? Observer::self() : 0L;
|
|
switch (state) {
|
|
case STATE_COPYING_FILES:
|
|
emit processedFiles( this, m_processedFiles );
|
|
if (observer) observer->slotProcessedFiles(this, m_processedFiles);
|
|
if (d->m_bURLDirty)
|
|
{
|
|
// Only emit urls when they changed. This saves time, and fixes #66281
|
|
d->m_bURLDirty = false;
|
|
if (m_mode==Move)
|
|
{
|
|
if (observer) observer->slotMoving( this, m_currentSrcURL, m_currentDestURL);
|
|
emit moving( this, m_currentSrcURL, m_currentDestURL);
|
|
}
|
|
else if (m_mode==Link)
|
|
{
|
|
if (observer) observer->slotCopying( this, m_currentSrcURL, m_currentDestURL ); // we don't have a slotLinking
|
|
emit linking( this, m_currentSrcURL.path(), m_currentDestURL );
|
|
}
|
|
else
|
|
{
|
|
if (observer) observer->slotCopying( this, m_currentSrcURL, m_currentDestURL );
|
|
emit copying( this, m_currentSrcURL, m_currentDestURL );
|
|
}
|
|
}
|
|
break;
|
|
|
|
case STATE_CREATING_DIRS:
|
|
if (observer) observer->slotProcessedDirs( this, m_processedDirs );
|
|
emit processedDirs( this, m_processedDirs );
|
|
if (d->m_bURLDirty)
|
|
{
|
|
d->m_bURLDirty = false;
|
|
emit creatingDir( this, m_currentDestURL );
|
|
if (observer) observer->slotCreatingDir( this, m_currentDestURL);
|
|
}
|
|
break;
|
|
|
|
case STATE_STATING:
|
|
case STATE_LISTING:
|
|
if (d->m_bURLDirty)
|
|
{
|
|
d->m_bURLDirty = false;
|
|
if (observer) observer->slotCopying( this, m_currentSrcURL, m_currentDestURL );
|
|
}
|
|
emit totalSize( this, m_totalSize );
|
|
emit totalFiles( this, files.count() );
|
|
emit totalDirs( this, dirs.count() );
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
void CopyJob::slotEntries(KIO::Job* job, const UDSEntryList& list)
|
|
{
|
|
UDSEntryListConstIterator it = list.begin();
|
|
UDSEntryListConstIterator end = list.end();
|
|
for (; it != end; ++it) {
|
|
UDSEntry::ConstIterator it2 = (*it).begin();
|
|
struct CopyInfo info;
|
|
info.permissions = -1;
|
|
info.mtime = (time_t) -1;
|
|
info.ctime = (time_t) -1;
|
|
info.size = (KIO::filesize_t)-1;
|
|
TQString displayName;
|
|
KURL url;
|
|
TQString localPath;
|
|
bool isDir = false;
|
|
for( ; it2 != (*it).end(); it2++ ) {
|
|
switch ((*it2).m_uds) {
|
|
case UDS_FILE_TYPE:
|
|
//info.type = (mode_t)((*it2).m_long);
|
|
isDir = S_ISDIR( (mode_t)((*it2).m_long) );
|
|
break;
|
|
case UDS_NAME: // recursive listing, displayName can be a/b/c/d
|
|
displayName = (*it2).m_str;
|
|
break;
|
|
case UDS_URL: // optional
|
|
url = KURL((*it2).m_str);
|
|
break;
|
|
case UDS_LOCAL_PATH:
|
|
localPath = (*it2).m_str;
|
|
break;
|
|
case UDS_LINK_DEST:
|
|
info.linkDest = (*it2).m_str;
|
|
break;
|
|
case UDS_ACCESS:
|
|
info.permissions = ((*it2).m_long);
|
|
break;
|
|
case UDS_SIZE:
|
|
info.size = (KIO::filesize_t)((*it2).m_long);
|
|
m_totalSize += info.size;
|
|
break;
|
|
case UDS_MODIFICATION_TIME:
|
|
info.mtime = (time_t)((*it2).m_long);
|
|
break;
|
|
case UDS_CREATION_TIME:
|
|
info.ctime = (time_t)((*it2).m_long);
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
if (displayName != ".." && displayName != ".")
|
|
{
|
|
bool hasCustomURL = !url.isEmpty() || !localPath.isEmpty();
|
|
if( !hasCustomURL ) {
|
|
// Make URL from displayName
|
|
url = ((SimpleJob *)job)->url();
|
|
if ( m_bCurrentSrcIsDir ) { // Only if src is a directory. Otherwise uSource is fine as is
|
|
//kdDebug(7007) << "adding path " << displayName << endl;
|
|
url.addPath( displayName );
|
|
}
|
|
}
|
|
//kdDebug(7007) << "displayName=" << displayName << " url=" << url << endl;
|
|
if (!localPath.isEmpty() && kio_resolve_local_urls) {
|
|
url = KURL();
|
|
url.setPath(localPath);
|
|
}
|
|
|
|
info.uSource = url;
|
|
info.uDest = m_currentDest;
|
|
//kdDebug(7007) << " uSource=" << info.uSource << " uDest(1)=" << info.uDest << endl;
|
|
// Append filename or dirname to destination URL, if allowed
|
|
if ( destinationState == DEST_IS_DIR &&
|
|
// "copy/move as <foo>" means 'foo' is the dest for the base srcurl
|
|
// (passed here during stating) but not its children (during listing)
|
|
( ! ( m_asMethod && state == STATE_STATING ) ) )
|
|
{
|
|
TQString destFileName;
|
|
if ( hasCustomURL &&
|
|
KProtocolInfo::fileNameUsedForCopying( url ) == KProtocolInfo::FromURL ) {
|
|
//destFileName = url.fileName(); // Doesn't work for recursive listing
|
|
// Count the number of prefixes used by the recursive listjob
|
|
int numberOfSlashes = displayName.contains( '/' ); // don't make this a find()!
|
|
TQString path = url.path();
|
|
int pos = 0;
|
|
for ( int n = 0; n < numberOfSlashes + 1; ++n ) {
|
|
pos = path.findRev( '/', pos - 1 );
|
|
if ( pos == -1 ) { // error
|
|
kdWarning(7007) << "kioslave bug: not enough slashes in UDS_URL " << path << " - looking for " << numberOfSlashes << " slashes" << endl;
|
|
break;
|
|
}
|
|
}
|
|
if ( pos >= 0 ) {
|
|
destFileName = path.mid( pos + 1 );
|
|
}
|
|
|
|
} else { // destination filename taken from UDS_NAME
|
|
destFileName = displayName;
|
|
}
|
|
|
|
// Here we _really_ have to add some filename to the dest.
|
|
// Otherwise, we end up with e.g. dest=..../Desktop/ itself.
|
|
// (This can happen when dropping a link to a webpage with no path)
|
|
if ( destFileName.isEmpty() )
|
|
destFileName = KIO::encodeFileName( info.uSource.prettyURL() );
|
|
|
|
//kdDebug(7007) << " adding destFileName=" << destFileName << endl;
|
|
info.uDest.addPath( destFileName );
|
|
}
|
|
//kdDebug(7007) << " uDest(2)=" << info.uDest << endl;
|
|
//kdDebug(7007) << " " << info.uSource << " -> " << info.uDest << endl;
|
|
if ( info.linkDest.isEmpty() && isDir && m_mode != Link ) // Dir
|
|
{
|
|
dirs.append( info ); // Directories
|
|
if (m_mode == Move)
|
|
dirsToRemove.append( info.uSource );
|
|
}
|
|
else {
|
|
files.append( info ); // Files and any symlinks
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
void CopyJob::skipSrc()
|
|
{
|
|
m_dest = d->m_globalDest;
|
|
destinationState = d->m_globalDestinationState;
|
|
++m_currentStatSrc;
|
|
skip( m_currentSrcURL );
|
|
statCurrentSrc();
|
|
}
|
|
|
|
void CopyJob::statNextSrc()
|
|
{
|
|
/* Revert to the global destination, the one that applies to all source urls.
|
|
* Imagine you copy the items a b and c into /d, but /d/b exists so the user uses "Rename" to put it in /foo/b instead.
|
|
* m_dest is /foo/b for b, but we have to revert to /d for item c and following.
|
|
*/
|
|
m_dest = d->m_globalDest;
|
|
destinationState = d->m_globalDestinationState;
|
|
++m_currentStatSrc;
|
|
statCurrentSrc();
|
|
}
|
|
|
|
void CopyJob::statCurrentSrc()
|
|
{
|
|
if ( m_currentStatSrc != m_srcList.end() )
|
|
{
|
|
m_currentSrcURL = (*m_currentStatSrc);
|
|
d->m_bURLDirty = true;
|
|
if ( m_mode == Link )
|
|
{
|
|
// Skip the "stating the source" stage, we don't need it for linking
|
|
m_currentDest = m_dest;
|
|
struct CopyInfo info;
|
|
info.permissions = -1;
|
|
info.mtime = (time_t) -1;
|
|
info.ctime = (time_t) -1;
|
|
info.size = (KIO::filesize_t)-1;
|
|
info.uSource = m_currentSrcURL;
|
|
info.uDest = m_currentDest;
|
|
// Append filename or dirname to destination URL, if allowed
|
|
if ( destinationState == DEST_IS_DIR && !m_asMethod )
|
|
{
|
|
if (
|
|
(m_currentSrcURL.protocol() == info.uDest.protocol()) &&
|
|
(m_currentSrcURL.host() == info.uDest.host()) &&
|
|
(m_currentSrcURL.port() == info.uDest.port()) &&
|
|
(m_currentSrcURL.user() == info.uDest.user()) &&
|
|
(m_currentSrcURL.pass() == info.uDest.pass()) )
|
|
{
|
|
// This is the case of creating a real symlink
|
|
info.uDest.addPath( m_currentSrcURL.fileName() );
|
|
}
|
|
else
|
|
{
|
|
// Different protocols, we'll create a .desktop file
|
|
// We have to change the extension anyway, so while we're at it,
|
|
// name the file like the URL
|
|
info.uDest.addPath( KIO::encodeFileName( m_currentSrcURL.prettyURL() )+".desktop" );
|
|
}
|
|
}
|
|
files.append( info ); // Files and any symlinks
|
|
statNextSrc(); // we could use a loop instead of a recursive call :)
|
|
return;
|
|
}
|
|
else if ( m_mode == Move && (
|
|
// Don't go renaming right away if we need a stat() to find out the destination filename
|
|
KProtocolInfo::fileNameUsedForCopying( m_currentSrcURL ) == KProtocolInfo::FromURL ||
|
|
destinationState != DEST_IS_DIR || m_asMethod )
|
|
)
|
|
{
|
|
// If moving, before going for the full stat+[list+]copy+del thing, try to rename
|
|
// The logic is pretty similar to FileCopyJob::slotStart()
|
|
if ( (m_currentSrcURL.protocol() == m_dest.protocol()) &&
|
|
(m_currentSrcURL.host() == m_dest.host()) &&
|
|
(m_currentSrcURL.port() == m_dest.port()) &&
|
|
(m_currentSrcURL.user() == m_dest.user()) &&
|
|
(m_currentSrcURL.pass() == m_dest.pass()) )
|
|
{
|
|
startRenameJob( m_currentSrcURL );
|
|
return;
|
|
}
|
|
else if ( m_currentSrcURL.isLocalFile() && KProtocolInfo::canRenameFromFile( m_dest ) )
|
|
{
|
|
startRenameJob( m_dest );
|
|
return;
|
|
}
|
|
else if ( m_dest.isLocalFile() && KProtocolInfo::canRenameToFile( m_currentSrcURL ) )
|
|
{
|
|
startRenameJob( m_currentSrcURL );
|
|
return;
|
|
}
|
|
}
|
|
|
|
// if the file system doesn't support deleting, we do not even stat
|
|
if (m_mode == Move && !KProtocolInfo::supportsDeleting(m_currentSrcURL)) {
|
|
TQGuardedPtr<CopyJob> that = this;
|
|
if (isInteractive())
|
|
KMessageBox::information( 0, buildErrorString(ERR_CANNOT_DELETE, m_currentSrcURL.prettyURL()));
|
|
if (that)
|
|
statNextSrc(); // we could use a loop instead of a recursive call :)
|
|
return;
|
|
}
|
|
|
|
// Stat the next src url
|
|
Job * job = KIO::stat( m_currentSrcURL, true, 2, false );
|
|
//kdDebug(7007) << "KIO::stat on " << m_currentSrcURL << endl;
|
|
state = STATE_STATING;
|
|
addSubjob(job);
|
|
m_currentDestURL=m_dest;
|
|
m_bOnlyRenames = false;
|
|
d->m_bURLDirty = true;
|
|
}
|
|
else
|
|
{
|
|
// Finished the stat'ing phase
|
|
// First make sure that the totals were correctly emitted
|
|
state = STATE_STATING;
|
|
d->m_bURLDirty = true;
|
|
slotReport();
|
|
if (!dirs.isEmpty())
|
|
emit aboutToCreate( this, dirs );
|
|
if (!files.isEmpty())
|
|
emit aboutToCreate( this, files );
|
|
// Check if we are copying a single file
|
|
m_bSingleFileCopy = ( files.count() == 1 && dirs.isEmpty() );
|
|
// Then start copying things
|
|
state = STATE_CREATING_DIRS;
|
|
createNextDir();
|
|
}
|
|
}
|
|
|
|
void CopyJob::startRenameJob( const KURL& slave_url )
|
|
{
|
|
KURL dest = m_dest;
|
|
// Append filename or dirname to destination URL, if allowed
|
|
if ( destinationState == DEST_IS_DIR && !m_asMethod )
|
|
dest.addPath( m_currentSrcURL.fileName() );
|
|
kdDebug(7007) << "This seems to be a suitable case for trying to rename before stat+[list+]copy+del" << endl;
|
|
state = STATE_RENAMING;
|
|
|
|
struct CopyInfo info;
|
|
info.permissions = -1;
|
|
info.mtime = (time_t) -1;
|
|
info.ctime = (time_t) -1;
|
|
info.size = (KIO::filesize_t)-1;
|
|
info.uSource = m_currentSrcURL;
|
|
info.uDest = dest;
|
|
TQValueList<CopyInfo> files;
|
|
files.append(info);
|
|
emit aboutToCreate( this, files );
|
|
|
|
KIO_ARGS << m_currentSrcURL << dest << (TQ_INT8) false /*no overwrite*/;
|
|
SimpleJob * newJob = new SimpleJob(slave_url, CMD_RENAME, packedArgs, false);
|
|
Scheduler::scheduleJob(newJob);
|
|
addSubjob( newJob );
|
|
if ( m_currentSrcURL.directory() != dest.directory() ) // For the user, moving isn't renaming. Only renaming is.
|
|
m_bOnlyRenames = false;
|
|
}
|
|
|
|
void CopyJob::startListing( const KURL & src )
|
|
{
|
|
state = STATE_LISTING;
|
|
d->m_bURLDirty = true;
|
|
ListJob * newjob = listRecursive( src, false );
|
|
newjob->setUnrestricted(true);
|
|
connect(newjob, TQT_SIGNAL(entries( KIO::Job *,
|
|
const KIO::UDSEntryList& )),
|
|
TQT_SLOT( slotEntries( KIO::Job*,
|
|
const KIO::UDSEntryList& )));
|
|
addSubjob( newjob );
|
|
}
|
|
|
|
void CopyJob::skip( const KURL & sourceUrl )
|
|
{
|
|
// Check if this is one if toplevel sources
|
|
// If yes, remove it from m_srcList, for a correct FilesRemoved() signal
|
|
//kdDebug(7007) << "CopyJob::skip: looking for " << sourceUrl << endl;
|
|
KURL::List::Iterator sit = m_srcList.find( sourceUrl );
|
|
if ( sit != m_srcList.end() )
|
|
{
|
|
//kdDebug(7007) << "CopyJob::skip: removing " << sourceUrl << " from list" << endl;
|
|
m_srcList.remove( sit );
|
|
}
|
|
dirsToRemove.remove( sourceUrl );
|
|
}
|
|
|
|
bool CopyJob::shouldOverwrite( const TQString& path ) const
|
|
{
|
|
if ( m_bOverwriteAll )
|
|
return true;
|
|
TQStringList::ConstIterator sit = m_overwriteList.begin();
|
|
for( ; sit != m_overwriteList.end(); ++sit )
|
|
if ( path.startsWith( *sit ) )
|
|
return true;
|
|
return false;
|
|
}
|
|
|
|
bool CopyJob::shouldSkip( const TQString& path ) const
|
|
{
|
|
TQStringList::ConstIterator sit = m_skipList.begin();
|
|
for( ; sit != m_skipList.end(); ++sit )
|
|
if ( path.startsWith( *sit ) )
|
|
return true;
|
|
return false;
|
|
}
|
|
|
|
void CopyJob::slotResultCreatingDirs( Job * job )
|
|
{
|
|
// The dir we are trying to create:
|
|
TQValueList<CopyInfo>::Iterator it = dirs.begin();
|
|
// Was there an error creating a dir ?
|
|
if ( job->error() )
|
|
{
|
|
m_conflictError = job->error();
|
|
if ( (m_conflictError == ERR_DIR_ALREADY_EXIST)
|
|
|| (m_conflictError == ERR_FILE_ALREADY_EXIST) ) // can't happen?
|
|
{
|
|
KURL oldURL = ((SimpleJob*)job)->url();
|
|
// Should we skip automatically ?
|
|
if ( m_bAutoSkip ) {
|
|
// We don't want to copy files in this directory, so we put it on the skip list
|
|
m_skipList.append( oldURL.path( 1 ) );
|
|
skip( oldURL );
|
|
dirs.remove( it ); // Move on to next dir
|
|
} else {
|
|
// Did the user choose to overwrite already?
|
|
const TQString destFile = (*it).uDest.path();
|
|
if ( shouldOverwrite( destFile ) ) { // overwrite => just skip
|
|
emit copyingDone( this, ( *it ).uSource, ( *it ).uDest, true /* directory */, false /* renamed */ );
|
|
dirs.remove( it ); // Move on to next dir
|
|
} else {
|
|
if ( !isInteractive() ) {
|
|
Job::slotResult( job ); // will set the error and emit result(this)
|
|
return;
|
|
}
|
|
|
|
assert( ((SimpleJob*)job)->url().url() == (*it).uDest.url() );
|
|
subjobs.remove( job );
|
|
assert ( subjobs.isEmpty() ); // We should have only one job at a time ...
|
|
|
|
// We need to stat the existing dir, to get its last-modification time
|
|
KURL existingDest( (*it).uDest );
|
|
SimpleJob * newJob = KIO::stat( existingDest, false, 2, false );
|
|
Scheduler::scheduleJob(newJob);
|
|
kdDebug(7007) << "KIO::stat for resolving conflict on " << existingDest << endl;
|
|
state = STATE_CONFLICT_CREATING_DIRS;
|
|
addSubjob(newJob);
|
|
return; // Don't move to next dir yet !
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
// Severe error, abort
|
|
Job::slotResult( job ); // will set the error and emit result(this)
|
|
return;
|
|
}
|
|
}
|
|
else // no error : remove from list, to move on to next dir
|
|
{
|
|
//this is required for the undo feature
|
|
emit copyingDone( this, (*it).uSource, (*it).uDest, true, false );
|
|
d->m_directoriesCopied.append( *it );
|
|
dirs.remove( it );
|
|
}
|
|
|
|
m_processedDirs++;
|
|
//emit processedDirs( this, m_processedDirs );
|
|
subjobs.remove( job );
|
|
assert( subjobs.isEmpty() ); // We should have only one job at a time ...
|
|
createNextDir();
|
|
}
|
|
|
|
void CopyJob::slotResultConflictCreatingDirs( KIO::Job * job )
|
|
{
|
|
// We come here after a conflict has been detected and we've stated the existing dir
|
|
|
|
// The dir we were trying to create:
|
|
TQValueList<CopyInfo>::Iterator it = dirs.begin();
|
|
// Its modification time:
|
|
time_t destmtime = (time_t)-1;
|
|
time_t destctime = (time_t)-1;
|
|
KIO::filesize_t destsize = 0;
|
|
TQString linkDest;
|
|
|
|
UDSEntry entry = ((KIO::StatJob*)job)->statResult();
|
|
KIO::UDSEntry::ConstIterator it2 = entry.begin();
|
|
for( ; it2 != entry.end(); it2++ ) {
|
|
switch ((*it2).m_uds) {
|
|
case UDS_MODIFICATION_TIME:
|
|
destmtime = (time_t)((*it2).m_long);
|
|
break;
|
|
case UDS_CREATION_TIME:
|
|
destctime = (time_t)((*it2).m_long);
|
|
break;
|
|
case UDS_SIZE:
|
|
destsize = (*it2).m_long;
|
|
break;
|
|
case UDS_LINK_DEST:
|
|
linkDest = (*it2).m_str;
|
|
break;
|
|
}
|
|
}
|
|
subjobs.remove( job );
|
|
assert ( subjobs.isEmpty() ); // We should have only one job at a time ...
|
|
|
|
// Always multi and skip (since there are files after that)
|
|
RenameDlg_Mode mode = (RenameDlg_Mode)( M_MULTI | M_SKIP );
|
|
// Overwrite only if the existing thing is a dir (no chance with a file)
|
|
if ( m_conflictError == ERR_DIR_ALREADY_EXIST )
|
|
{
|
|
if( (*it).uSource == (*it).uDest ||
|
|
((*it).uSource.protocol() == (*it).uDest.protocol() &&
|
|
(*it).uSource.path(-1) == linkDest) )
|
|
mode = (RenameDlg_Mode)( mode | M_OVERWRITE_ITSELF);
|
|
else
|
|
mode = (RenameDlg_Mode)( mode | M_OVERWRITE );
|
|
}
|
|
|
|
TQString existingDest = (*it).uDest.path();
|
|
TQString newPath;
|
|
if (m_reportTimer)
|
|
m_reportTimer->stop();
|
|
RenameDlg_Result r = Observer::self()->open_RenameDlg( this, i18n("Folder Already Exists"),
|
|
(*it).uSource.url(),
|
|
(*it).uDest.url(),
|
|
mode, newPath,
|
|
(*it).size, destsize,
|
|
(*it).ctime, destctime,
|
|
(*it).mtime, destmtime );
|
|
if (m_reportTimer)
|
|
m_reportTimer->start(REPORT_TIMEOUT,false);
|
|
switch ( r ) {
|
|
case R_CANCEL:
|
|
m_error = ERR_USER_CANCELED;
|
|
emitResult();
|
|
return;
|
|
case R_RENAME:
|
|
{
|
|
TQString oldPath = (*it).uDest.path( 1 );
|
|
KURL newUrl( (*it).uDest );
|
|
newUrl.setPath( newPath );
|
|
emit renamed( this, (*it).uDest, newUrl ); // for e.g. kpropsdlg
|
|
|
|
// Change the current one and strip the trailing '/'
|
|
(*it).uDest.setPath( newUrl.path( -1 ) );
|
|
newPath = newUrl.path( 1 ); // With trailing slash
|
|
TQValueList<CopyInfo>::Iterator renamedirit = it;
|
|
++renamedirit;
|
|
// Change the name of subdirectories inside the directory
|
|
for( ; renamedirit != dirs.end() ; ++renamedirit )
|
|
{
|
|
TQString path = (*renamedirit).uDest.path();
|
|
if ( path.left(oldPath.length()) == oldPath ) {
|
|
TQString n = path;
|
|
n.replace( 0, oldPath.length(), newPath );
|
|
kdDebug(7007) << "dirs list: " << (*renamedirit).uSource.path()
|
|
<< " was going to be " << path
|
|
<< ", changed into " << n << endl;
|
|
(*renamedirit).uDest.setPath( n );
|
|
}
|
|
}
|
|
// Change filenames inside the directory
|
|
TQValueList<CopyInfo>::Iterator renamefileit = files.begin();
|
|
for( ; renamefileit != files.end() ; ++renamefileit )
|
|
{
|
|
TQString path = (*renamefileit).uDest.path();
|
|
if ( path.left(oldPath.length()) == oldPath ) {
|
|
TQString n = path;
|
|
n.replace( 0, oldPath.length(), newPath );
|
|
kdDebug(7007) << "files list: " << (*renamefileit).uSource.path()
|
|
<< " was going to be " << path
|
|
<< ", changed into " << n << endl;
|
|
(*renamefileit).uDest.setPath( n );
|
|
}
|
|
}
|
|
if (!dirs.isEmpty())
|
|
emit aboutToCreate( this, dirs );
|
|
if (!files.isEmpty())
|
|
emit aboutToCreate( this, files );
|
|
}
|
|
break;
|
|
case R_AUTO_SKIP:
|
|
m_bAutoSkip = true;
|
|
// fall through
|
|
case R_SKIP:
|
|
m_skipList.append( existingDest );
|
|
skip( (*it).uSource );
|
|
// Move on to next dir
|
|
dirs.remove( it );
|
|
m_processedDirs++;
|
|
break;
|
|
case R_OVERWRITE:
|
|
m_overwriteList.append( existingDest );
|
|
emit copyingDone( this, ( *it ).uSource, ( *it ).uDest, true /* directory */, false /* renamed */ );
|
|
// Move on to next dir
|
|
dirs.remove( it );
|
|
m_processedDirs++;
|
|
break;
|
|
case R_OVERWRITE_ALL:
|
|
m_bOverwriteAll = true;
|
|
emit copyingDone( this, ( *it ).uSource, ( *it ).uDest, true /* directory */, false /* renamed */ );
|
|
// Move on to next dir
|
|
dirs.remove( it );
|
|
m_processedDirs++;
|
|
break;
|
|
default:
|
|
assert( 0 );
|
|
}
|
|
state = STATE_CREATING_DIRS;
|
|
//emit processedDirs( this, m_processedDirs );
|
|
createNextDir();
|
|
}
|
|
|
|
void CopyJob::createNextDir()
|
|
{
|
|
KURL udir;
|
|
if ( !dirs.isEmpty() )
|
|
{
|
|
// Take first dir to create out of list
|
|
TQValueList<CopyInfo>::Iterator it = dirs.begin();
|
|
// Is this URL on the skip list or the overwrite list ?
|
|
while( it != dirs.end() && udir.isEmpty() )
|
|
{
|
|
const TQString dir = (*it).uDest.path();
|
|
if ( shouldSkip( dir ) ) {
|
|
dirs.remove( it );
|
|
it = dirs.begin();
|
|
} else
|
|
udir = (*it).uDest;
|
|
}
|
|
}
|
|
if ( !udir.isEmpty() ) // any dir to create, finally ?
|
|
{
|
|
// Create the directory - with default permissions so that we can put files into it
|
|
// TODO : change permissions once all is finished; but for stuff coming from CDROM it sucks...
|
|
KIO::SimpleJob *newjob = KIO::mkdir( udir, -1 );
|
|
Scheduler::scheduleJob(newjob);
|
|
|
|
m_currentDestURL = udir;
|
|
d->m_bURLDirty = true;
|
|
|
|
addSubjob(newjob);
|
|
return;
|
|
}
|
|
else // we have finished creating dirs
|
|
{
|
|
emit processedDirs( this, m_processedDirs ); // make sure final number appears
|
|
if (m_progressId) Observer::self()->slotProcessedDirs( this, m_processedDirs );
|
|
|
|
state = STATE_COPYING_FILES;
|
|
m_processedFiles++; // Ralf wants it to start at 1, not 0
|
|
copyNextFile();
|
|
}
|
|
}
|
|
|
|
void CopyJob::slotResultCopyingFiles( Job * job )
|
|
{
|
|
// The file we were trying to copy:
|
|
TQValueList<CopyInfo>::Iterator it = files.begin();
|
|
if ( job->error() )
|
|
{
|
|
// Should we skip automatically ?
|
|
if ( m_bAutoSkip )
|
|
{
|
|
skip( (*it).uSource );
|
|
m_fileProcessedSize = (*it).size;
|
|
files.remove( it ); // Move on to next file
|
|
}
|
|
else
|
|
{
|
|
if ( !isInteractive() ) {
|
|
Job::slotResult( job ); // will set the error and emit result(this)
|
|
return;
|
|
}
|
|
|
|
m_conflictError = job->error(); // save for later
|
|
// Existing dest ?
|
|
if ( ( m_conflictError == ERR_FILE_ALREADY_EXIST )
|
|
|| ( m_conflictError == ERR_DIR_ALREADY_EXIST )
|
|
|| ( m_conflictError == ERR_IDENTICAL_FILES ) )
|
|
{
|
|
subjobs.remove( job );
|
|
assert ( subjobs.isEmpty() );
|
|
// We need to stat the existing file, to get its last-modification time
|
|
KURL existingFile( (*it).uDest );
|
|
SimpleJob * newJob = KIO::stat( existingFile, false, 2, false );
|
|
Scheduler::scheduleJob(newJob);
|
|
kdDebug(7007) << "KIO::stat for resolving conflict on " << existingFile << endl;
|
|
state = STATE_CONFLICT_COPYING_FILES;
|
|
addSubjob(newJob);
|
|
return; // Don't move to next file yet !
|
|
}
|
|
else
|
|
{
|
|
if ( m_bCurrentOperationIsLink && ::tqqt_cast<KIO::DeleteJob*>( job ) )
|
|
{
|
|
// Very special case, see a few lines below
|
|
// We are deleting the source of a symlink we successfully moved... ignore error
|
|
m_fileProcessedSize = (*it).size;
|
|
files.remove( it );
|
|
} else {
|
|
// Go directly to the conflict resolution, there is nothing to stat
|
|
slotResultConflictCopyingFiles( job );
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
} else // no error
|
|
{
|
|
// Special case for moving links. That operation needs two jobs, unlike others.
|
|
if ( m_bCurrentOperationIsLink && m_mode == Move
|
|
&& !::tqqt_cast<KIO::DeleteJob *>( job ) // Deleting source not already done
|
|
)
|
|
{
|
|
subjobs.remove( job );
|
|
assert ( subjobs.isEmpty() );
|
|
// The only problem with this trick is that the error handling for this del operation
|
|
// is not going to be right... see 'Very special case' above.
|
|
KIO::Job * newjob = KIO::del( (*it).uSource, false /*don't shred*/, false /*no GUI*/ );
|
|
addSubjob( newjob );
|
|
return; // Don't move to next file yet !
|
|
}
|
|
|
|
if ( m_bCurrentOperationIsLink )
|
|
{
|
|
TQString target = ( m_mode == Link ? (*it).uSource.path() : (*it).linkDest );
|
|
//required for the undo feature
|
|
emit copyingLinkDone( this, (*it).uSource, target, (*it).uDest );
|
|
}
|
|
else
|
|
//required for the undo feature
|
|
emit copyingDone( this, (*it).uSource, (*it).uDest, false, false );
|
|
// remove from list, to move on to next file
|
|
files.remove( it );
|
|
}
|
|
m_processedFiles++;
|
|
|
|
// clear processed size for last file and add it to overall processed size
|
|
m_processedSize += m_fileProcessedSize;
|
|
m_fileProcessedSize = 0;
|
|
|
|
//kdDebug(7007) << files.count() << " files remaining" << endl;
|
|
|
|
removeSubjob( job, true, false ); // merge metadata
|
|
assert ( subjobs.isEmpty() ); // We should have only one job at a time ...
|
|
copyNextFile();
|
|
}
|
|
|
|
void CopyJob::slotResultConflictCopyingFiles( KIO::Job * job )
|
|
{
|
|
// We come here after a conflict has been detected and we've stated the existing file
|
|
// The file we were trying to create:
|
|
TQValueList<CopyInfo>::Iterator it = files.begin();
|
|
|
|
RenameDlg_Result res;
|
|
TQString newPath;
|
|
|
|
if (m_reportTimer)
|
|
m_reportTimer->stop();
|
|
|
|
if ( ( m_conflictError == ERR_FILE_ALREADY_EXIST )
|
|
|| ( m_conflictError == ERR_DIR_ALREADY_EXIST )
|
|
|| ( m_conflictError == ERR_IDENTICAL_FILES ) )
|
|
{
|
|
// Its modification time:
|
|
time_t destmtime = (time_t)-1;
|
|
time_t destctime = (time_t)-1;
|
|
KIO::filesize_t destsize = 0;
|
|
TQString linkDest;
|
|
UDSEntry entry = ((KIO::StatJob*)job)->statResult();
|
|
KIO::UDSEntry::ConstIterator it2 = entry.begin();
|
|
for( ; it2 != entry.end(); it2++ ) {
|
|
switch ((*it2).m_uds) {
|
|
case UDS_MODIFICATION_TIME:
|
|
destmtime = (time_t)((*it2).m_long);
|
|
break;
|
|
case UDS_CREATION_TIME:
|
|
destctime = (time_t)((*it2).m_long);
|
|
break;
|
|
case UDS_SIZE:
|
|
destsize = (*it2).m_long;
|
|
break;
|
|
case UDS_LINK_DEST:
|
|
linkDest = (*it2).m_str;
|
|
break;
|
|
}
|
|
}
|
|
|
|
// Offer overwrite only if the existing thing is a file
|
|
// If src==dest, use "overwrite-itself"
|
|
RenameDlg_Mode mode;
|
|
bool isDir = true;
|
|
|
|
if( m_conflictError == ERR_DIR_ALREADY_EXIST )
|
|
mode = (RenameDlg_Mode) 0;
|
|
else
|
|
{
|
|
if ( (*it).uSource == (*it).uDest ||
|
|
((*it).uSource.protocol() == (*it).uDest.protocol() &&
|
|
(*it).uSource.path(-1) == linkDest) )
|
|
mode = M_OVERWRITE_ITSELF;
|
|
else
|
|
mode = M_OVERWRITE;
|
|
isDir = false;
|
|
}
|
|
|
|
if ( m_bSingleFileCopy )
|
|
mode = (RenameDlg_Mode) ( mode | M_SINGLE );
|
|
else
|
|
mode = (RenameDlg_Mode) ( mode | M_MULTI | M_SKIP );
|
|
|
|
res = Observer::self()->open_RenameDlg( this, !isDir ?
|
|
i18n("File Already Exists") : i18n("Already Exists as Folder"),
|
|
(*it).uSource.url(),
|
|
(*it).uDest.url(),
|
|
mode, newPath,
|
|
(*it).size, destsize,
|
|
(*it).ctime, destctime,
|
|
(*it).mtime, destmtime );
|
|
|
|
}
|
|
else
|
|
{
|
|
if ( job->error() == ERR_USER_CANCELED )
|
|
res = R_CANCEL;
|
|
else if ( !isInteractive() ) {
|
|
Job::slotResult( job ); // will set the error and emit result(this)
|
|
return;
|
|
}
|
|
else
|
|
{
|
|
SkipDlg_Result skipResult = Observer::self()->open_SkipDlg( this, files.count() > 1,
|
|
job->errorString() );
|
|
|
|
// Convert the return code from SkipDlg into a RenameDlg code
|
|
res = ( skipResult == S_SKIP ) ? R_SKIP :
|
|
( skipResult == S_AUTO_SKIP ) ? R_AUTO_SKIP :
|
|
R_CANCEL;
|
|
}
|
|
}
|
|
|
|
if (m_reportTimer)
|
|
m_reportTimer->start(REPORT_TIMEOUT,false);
|
|
|
|
subjobs.remove( job );
|
|
assert ( subjobs.isEmpty() );
|
|
switch ( res ) {
|
|
case R_CANCEL:
|
|
m_error = ERR_USER_CANCELED;
|
|
emitResult();
|
|
return;
|
|
case R_RENAME:
|
|
{
|
|
KURL newUrl( (*it).uDest );
|
|
newUrl.setPath( newPath );
|
|
emit renamed( this, (*it).uDest, newUrl ); // for e.g. kpropsdlg
|
|
(*it).uDest = newUrl;
|
|
|
|
TQValueList<CopyInfo> files;
|
|
files.append(*it);
|
|
emit aboutToCreate( this, files );
|
|
}
|
|
break;
|
|
case R_AUTO_SKIP:
|
|
m_bAutoSkip = true;
|
|
// fall through
|
|
case R_SKIP:
|
|
// Move on to next file
|
|
skip( (*it).uSource );
|
|
m_processedSize += (*it).size;
|
|
files.remove( it );
|
|
m_processedFiles++;
|
|
break;
|
|
case R_OVERWRITE_ALL:
|
|
m_bOverwriteAll = true;
|
|
break;
|
|
case R_OVERWRITE:
|
|
// Add to overwrite list, so that copyNextFile knows to overwrite
|
|
m_overwriteList.append( (*it).uDest.path() );
|
|
break;
|
|
default:
|
|
assert( 0 );
|
|
}
|
|
state = STATE_COPYING_FILES;
|
|
//emit processedFiles( this, m_processedFiles );
|
|
copyNextFile();
|
|
}
|
|
|
|
void CopyJob::copyNextFile()
|
|
{
|
|
bool bCopyFile = false;
|
|
//kdDebug(7007) << "CopyJob::copyNextFile()" << endl;
|
|
// Take the first file in the list
|
|
TQValueList<CopyInfo>::Iterator it = files.begin();
|
|
// Is this URL on the skip list ?
|
|
while (it != files.end() && !bCopyFile)
|
|
{
|
|
const TQString destFile = (*it).uDest.path();
|
|
bCopyFile = !shouldSkip( destFile );
|
|
if ( !bCopyFile ) {
|
|
files.remove( it );
|
|
it = files.begin();
|
|
}
|
|
}
|
|
|
|
if (bCopyFile) // any file to create, finally ?
|
|
{
|
|
// Do we set overwrite ?
|
|
bool bOverwrite;
|
|
const TQString destFile = (*it).uDest.path();
|
|
kdDebug(7007) << "copying " << destFile << endl;
|
|
if ( (*it).uDest == (*it).uSource )
|
|
bOverwrite = false;
|
|
else
|
|
bOverwrite = shouldOverwrite( destFile );
|
|
|
|
m_bCurrentOperationIsLink = false;
|
|
KIO::Job * newjob = 0L;
|
|
if ( m_mode == Link )
|
|
{
|
|
//kdDebug(7007) << "Linking" << endl;
|
|
if (
|
|
((*it).uSource.protocol() == (*it).uDest.protocol()) &&
|
|
((*it).uSource.host() == (*it).uDest.host()) &&
|
|
((*it).uSource.port() == (*it).uDest.port()) &&
|
|
((*it).uSource.user() == (*it).uDest.user()) &&
|
|
((*it).uSource.pass() == (*it).uDest.pass()) )
|
|
{
|
|
// This is the case of creating a real symlink
|
|
KIO::SimpleJob *newJob = KIO::symlink( (*it).uSource.path(), (*it).uDest, bOverwrite, false /*no GUI*/ );
|
|
newjob = newJob;
|
|
Scheduler::scheduleJob(newJob);
|
|
//kdDebug(7007) << "CopyJob::copyNextFile : Linking target=" << (*it).uSource.path() << " link=" << (*it).uDest << endl;
|
|
//emit linking( this, (*it).uSource.path(), (*it).uDest );
|
|
m_bCurrentOperationIsLink = true;
|
|
m_currentSrcURL=(*it).uSource;
|
|
m_currentDestURL=(*it).uDest;
|
|
d->m_bURLDirty = true;
|
|
//Observer::self()->slotCopying( this, (*it).uSource, (*it).uDest ); // should be slotLinking perhaps
|
|
} else {
|
|
//kdDebug(7007) << "CopyJob::copyNextFile : Linking URL=" << (*it).uSource << " link=" << (*it).uDest << endl;
|
|
if ( (*it).uDest.isLocalFile() )
|
|
{
|
|
bool devicesOk=false;
|
|
|
|
// if the source is a devices url, handle it a littlebit special
|
|
if ((*it).uSource.protocol()==TQString::tqfromLatin1("devices"))
|
|
{
|
|
TQByteArray data;
|
|
TQByteArray param;
|
|
TQCString retType;
|
|
TQDataStream streamout(param,IO_WriteOnly);
|
|
streamout<<(*it).uSource;
|
|
streamout<<(*it).uDest;
|
|
if ( kapp && kapp->dcopClient()->call( "kded",
|
|
"mountwatcher", "createLink(KURL, KURL)", param,retType,data,false ) )
|
|
{
|
|
TQDataStream streamin(data,IO_ReadOnly);
|
|
streamin>>devicesOk;
|
|
}
|
|
if (devicesOk)
|
|
{
|
|
files.remove( it );
|
|
m_processedFiles++;
|
|
//emit processedFiles( this, m_processedFiles );
|
|
copyNextFile();
|
|
return;
|
|
}
|
|
}
|
|
|
|
if (!devicesOk)
|
|
{
|
|
TQString path = (*it).uDest.path();
|
|
//kdDebug(7007) << "CopyJob::copyNextFile path=" << path << endl;
|
|
TQFile f( path );
|
|
if ( f.open( IO_ReadWrite ) )
|
|
{
|
|
f.close();
|
|
KSimpleConfig config( path );
|
|
config.setDesktopGroup();
|
|
KURL url = (*it).uSource;
|
|
url.setPass( "" );
|
|
config.writePathEntry( TQString::tqfromLatin1("URL"), url.url() );
|
|
config.writeEntry( TQString::tqfromLatin1("Name"), url.url() );
|
|
config.writeEntry( TQString::tqfromLatin1("Type"), TQString::tqfromLatin1("Link") );
|
|
TQString protocol = (*it).uSource.protocol();
|
|
if ( protocol == TQString::tqfromLatin1("ftp") )
|
|
config.writeEntry( TQString::tqfromLatin1("Icon"), TQString::tqfromLatin1("ftp") );
|
|
else if ( protocol == TQString::tqfromLatin1("http") )
|
|
config.writeEntry( TQString::tqfromLatin1("Icon"), TQString::tqfromLatin1("www") );
|
|
else if ( protocol == TQString::tqfromLatin1("info") )
|
|
config.writeEntry( TQString::tqfromLatin1("Icon"), TQString::tqfromLatin1("info") );
|
|
else if ( protocol == TQString::tqfromLatin1("mailto") ) // sven:
|
|
config.writeEntry( TQString::tqfromLatin1("Icon"), TQString::tqfromLatin1("kmail") ); // added mailto: support
|
|
else
|
|
config.writeEntry( TQString::tqfromLatin1("Icon"), TQString::tqfromLatin1("unknown") );
|
|
config.sync();
|
|
files.remove( it );
|
|
m_processedFiles++;
|
|
//emit processedFiles( this, m_processedFiles );
|
|
copyNextFile();
|
|
return;
|
|
}
|
|
else
|
|
{
|
|
kdDebug(7007) << "CopyJob::copyNextFile ERR_CANNOT_OPEN_FOR_WRITING" << endl;
|
|
m_error = ERR_CANNOT_OPEN_FOR_WRITING;
|
|
m_errorText = (*it).uDest.path();
|
|
emitResult();
|
|
return;
|
|
}
|
|
}
|
|
} else {
|
|
// Todo: not show "link" on remote dirs if the src urls are not from the same protocol+host+...
|
|
m_error = ERR_CANNOT_SYMLINK;
|
|
m_errorText = (*it).uDest.prettyURL();
|
|
emitResult();
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
else if ( !(*it).linkDest.isEmpty() &&
|
|
((*it).uSource.protocol() == (*it).uDest.protocol()) &&
|
|
((*it).uSource.host() == (*it).uDest.host()) &&
|
|
((*it).uSource.port() == (*it).uDest.port()) &&
|
|
((*it).uSource.user() == (*it).uDest.user()) &&
|
|
((*it).uSource.pass() == (*it).uDest.pass()))
|
|
// Copying a symlink - only on the same protocol/host/etc. (#5601, downloading an FTP file through its link),
|
|
{
|
|
KIO::SimpleJob *newJob = KIO::symlink( (*it).linkDest, (*it).uDest, bOverwrite, false /*no GUI*/ );
|
|
Scheduler::scheduleJob(newJob);
|
|
newjob = newJob;
|
|
//kdDebug(7007) << "CopyJob::copyNextFile : Linking target=" << (*it).linkDest << " link=" << (*it).uDest << endl;
|
|
//emit linking( this, (*it).linkDest, (*it).uDest );
|
|
m_currentSrcURL=(*it).linkDest;
|
|
m_currentDestURL=(*it).uDest;
|
|
d->m_bURLDirty = true;
|
|
//Observer::self()->slotCopying( this, (*it).linkDest, (*it).uDest ); // should be slotLinking perhaps
|
|
m_bCurrentOperationIsLink = true;
|
|
// NOTE: if we are moving stuff, the deletion of the source will be done in slotResultCopyingFiles
|
|
} else if (m_mode == Move) // Moving a file
|
|
{
|
|
KIO::FileCopyJob * moveJob = KIO::file_move( (*it).uSource, (*it).uDest, (*it).permissions, bOverwrite, false, false/*no GUI*/ );
|
|
moveJob->setSourceSize64( (*it).size );
|
|
newjob = moveJob;
|
|
//kdDebug(7007) << "CopyJob::copyNextFile : Moving " << (*it).uSource << " to " << (*it).uDest << endl;
|
|
//emit moving( this, (*it).uSource, (*it).uDest );
|
|
m_currentSrcURL=(*it).uSource;
|
|
m_currentDestURL=(*it).uDest;
|
|
d->m_bURLDirty = true;
|
|
//Observer::self()->slotMoving( this, (*it).uSource, (*it).uDest );
|
|
}
|
|
else // Copying a file
|
|
{
|
|
// If source isn't local and target is local, we ignore the original permissions
|
|
// Otherwise, files downloaded from HTTP end up with -r--r--r--
|
|
bool remoteSource = !KProtocolInfo::supportsListing((*it).uSource);
|
|
int permissions = (*it).permissions;
|
|
if ( d->m_defaultPermissions || ( remoteSource && (*it).uDest.isLocalFile() ) )
|
|
permissions = -1;
|
|
KIO::FileCopyJob * copyJob = KIO::file_copy( (*it).uSource, (*it).uDest, permissions, bOverwrite, false, false/*no GUI*/ );
|
|
copyJob->setParentJob( this ); // in case of rename dialog
|
|
copyJob->setSourceSize64( (*it).size );
|
|
copyJob->setModificationTime( (*it).mtime );
|
|
newjob = copyJob;
|
|
//kdDebug(7007) << "CopyJob::copyNextFile : Copying " << (*it).uSource << " to " << (*it).uDest << endl;
|
|
m_currentSrcURL=(*it).uSource;
|
|
m_currentDestURL=(*it).uDest;
|
|
d->m_bURLDirty = true;
|
|
}
|
|
addSubjob(newjob);
|
|
connect( newjob, TQT_SIGNAL( processedSize( KIO::Job*, KIO::filesize_t ) ),
|
|
this, TQT_SLOT( slotProcessedSize( KIO::Job*, KIO::filesize_t ) ) );
|
|
connect( newjob, TQT_SIGNAL( totalSize( KIO::Job*, KIO::filesize_t ) ),
|
|
this, TQT_SLOT( slotTotalSize( KIO::Job*, KIO::filesize_t ) ) );
|
|
}
|
|
else
|
|
{
|
|
// We're done
|
|
//kdDebug(7007) << "copyNextFile finished" << endl;
|
|
deleteNextDir();
|
|
}
|
|
}
|
|
|
|
void CopyJob::deleteNextDir()
|
|
{
|
|
if ( m_mode == Move && !dirsToRemove.isEmpty() ) // some dirs to delete ?
|
|
{
|
|
state = STATE_DELETING_DIRS;
|
|
d->m_bURLDirty = true;
|
|
// Take first dir to delete out of list - last ones first !
|
|
KURL::List::Iterator it = dirsToRemove.fromLast();
|
|
SimpleJob *job = KIO::rmdir( *it );
|
|
Scheduler::scheduleJob(job);
|
|
dirsToRemove.remove(it);
|
|
addSubjob( job );
|
|
}
|
|
else
|
|
{
|
|
// This step is done, move on
|
|
setNextDirAttribute();
|
|
}
|
|
}
|
|
|
|
void CopyJob::setNextDirAttribute()
|
|
{
|
|
if ( !d->m_directoriesCopied.isEmpty() )
|
|
{
|
|
state = STATE_SETTING_DIR_ATTRIBUTES;
|
|
#ifdef Q_OS_UNIX
|
|
// TODO KDE4: this should use a SlaveBase method, but we have none yet in KDE3.
|
|
TQValueList<CopyInfo>::Iterator it = d->m_directoriesCopied.begin();
|
|
for ( ; it != d->m_directoriesCopied.end() ; ++it ) {
|
|
const KURL& url = (*it).uDest;
|
|
if ( url.isLocalFile() && (*it).mtime != (time_t)-1 ) {
|
|
const TQCString path = TQFile::encodeName( url.path() );
|
|
KDE_struct_stat statbuf;
|
|
if (KDE_lstat(path, &statbuf) == 0) {
|
|
struct utimbuf utbuf;
|
|
utbuf.actime = statbuf.st_atime; // access time, unchanged
|
|
utbuf.modtime = (*it).mtime; // modification time
|
|
utime( path, &utbuf );
|
|
}
|
|
|
|
}
|
|
}
|
|
#endif
|
|
d->m_directoriesCopied.clear();
|
|
}
|
|
|
|
// No "else" here, since the above is a simple sync loop
|
|
|
|
{
|
|
// Finished - tell the world
|
|
if ( !m_bOnlyRenames )
|
|
{
|
|
KDirNotify_stub allDirNotify("*", "KDirNotify*");
|
|
KURL url( d->m_globalDest );
|
|
if ( d->m_globalDestinationState != DEST_IS_DIR || m_asMethod )
|
|
url.setPath( url.directory() );
|
|
//kdDebug(7007) << "KDirNotify'ing FilesAdded " << url << endl;
|
|
allDirNotify.FilesAdded( url );
|
|
|
|
if ( m_mode == Move && !m_srcList.isEmpty() ) {
|
|
//kdDebug(7007) << "KDirNotify'ing FilesRemoved " << m_srcList.toStringList() << endl;
|
|
allDirNotify.FilesRemoved( m_srcList );
|
|
}
|
|
}
|
|
if (m_reportTimer)
|
|
m_reportTimer->stop();
|
|
--m_processedFiles; // undo the "start at 1" hack
|
|
slotReport(); // display final numbers, important if progress dialog stays up
|
|
|
|
emitResult();
|
|
}
|
|
}
|
|
|
|
void CopyJob::slotProcessedSize( KIO::Job*, KIO::filesize_t data_size )
|
|
{
|
|
//kdDebug(7007) << "CopyJob::slotProcessedSize " << data_size << endl;
|
|
m_fileProcessedSize = data_size;
|
|
setProcessedSize(m_processedSize + m_fileProcessedSize);
|
|
|
|
if ( m_processedSize + m_fileProcessedSize > m_totalSize )
|
|
{
|
|
m_totalSize = m_processedSize + m_fileProcessedSize;
|
|
//kdDebug(7007) << "Adjusting m_totalSize to " << m_totalSize << endl;
|
|
emit totalSize( this, m_totalSize ); // safety
|
|
}
|
|
//kdDebug(7007) << "emit processedSize " << (unsigned long) (m_processedSize + m_fileProcessedSize) << endl;
|
|
emit processedSize( this, m_processedSize + m_fileProcessedSize );
|
|
emitPercent( m_processedSize + m_fileProcessedSize, m_totalSize );
|
|
}
|
|
|
|
void CopyJob::slotTotalSize( KIO::Job*, KIO::filesize_t size )
|
|
{
|
|
//kdDebug(7007) << "slotTotalSize: " << size << endl;
|
|
// Special case for copying a single file
|
|
// This is because some protocols don't implement stat properly
|
|
// (e.g. HTTP), and don't give us a size in some cases (redirection)
|
|
// so we'd rather rely on the size given for the transfer
|
|
if ( m_bSingleFileCopy && size > m_totalSize)
|
|
{
|
|
//kdDebug(7007) << "slotTotalSize: updating totalsize to " << size << endl;
|
|
m_totalSize = size;
|
|
emit totalSize( this, size );
|
|
}
|
|
}
|
|
|
|
void CopyJob::slotResultDeletingDirs( Job * job )
|
|
{
|
|
if (job->error())
|
|
{
|
|
// Couldn't remove directory. Well, perhaps it's not empty
|
|
// because the user pressed Skip for a given file in it.
|
|
// Let's not display "Could not remove dir ..." for each of those dir !
|
|
}
|
|
subjobs.remove( job );
|
|
assert ( subjobs.isEmpty() );
|
|
deleteNextDir();
|
|
}
|
|
|
|
#if 0 // TODO KDE4
|
|
void CopyJob::slotResultSettingDirAttributes( Job * job )
|
|
{
|
|
if (job->error())
|
|
{
|
|
// Couldn't set directory attributes. Ignore the error, it can happen
|
|
// with inferior file systems like VFAT.
|
|
// Let's not display warnings for each dir like "cp -a" does.
|
|
}
|
|
subjobs.remove( job );
|
|
assert ( subjobs.isEmpty() );
|
|
setNextDirAttribute();
|
|
}
|
|
#endif
|
|
|
|
void CopyJob::slotResultRenaming( Job* job )
|
|
{
|
|
int err = job->error();
|
|
const TQString errText = job->errorText();
|
|
removeSubjob( job, true, false ); // merge metadata
|
|
assert ( subjobs.isEmpty() );
|
|
// Determine dest again
|
|
KURL dest = m_dest;
|
|
if ( destinationState == DEST_IS_DIR && !m_asMethod )
|
|
dest.addPath( m_currentSrcURL.fileName() );
|
|
if ( err )
|
|
{
|
|
// Direct renaming didn't work. Try renaming to a temp name,
|
|
// this can help e.g. when renaming 'a' to 'A' on a VFAT partition.
|
|
// In that case it's the _same_ dir, we don't want to copy+del (data loss!)
|
|
if ( m_currentSrcURL.isLocalFile() && m_currentSrcURL.url(-1) != dest.url(-1) &&
|
|
m_currentSrcURL.url(-1).lower() == dest.url(-1).lower() &&
|
|
( err == ERR_FILE_ALREADY_EXIST ||
|
|
err == ERR_DIR_ALREADY_EXIST ||
|
|
err == ERR_IDENTICAL_FILES ) )
|
|
{
|
|
kdDebug(7007) << "Couldn't rename directly, dest already exists. Detected special case of lower/uppercase renaming in same dir, try with 2 rename calls" << endl;
|
|
TQCString _src( TQFile::encodeName(m_currentSrcURL.path()) );
|
|
TQCString _dest( TQFile::encodeName(dest.path()) );
|
|
KTempFile tmpFile( m_currentSrcURL.directory(false) );
|
|
TQCString _tmp( TQFile::encodeName(tmpFile.name()) );
|
|
kdDebug(7007) << "CopyJob::slotResult KTempFile status:" << tmpFile.status() << " using " << _tmp << " as intermediary" << endl;
|
|
tmpFile.unlink();
|
|
if ( ::rename( _src, _tmp ) == 0 )
|
|
{
|
|
if ( !TQFile::exists( _dest ) && ::rename( _tmp, _dest ) == 0 )
|
|
{
|
|
kdDebug(7007) << "Success." << endl;
|
|
err = 0;
|
|
}
|
|
else
|
|
{
|
|
// Revert back to original name!
|
|
if ( ::rename( _tmp, _src ) != 0 ) {
|
|
kdError(7007) << "Couldn't rename " << tmpFile.name() << " back to " << _src << " !" << endl;
|
|
// Severe error, abort
|
|
Job::slotResult( job ); // will set the error and emit result(this)
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
if ( err )
|
|
{
|
|
// This code is similar to CopyJob::slotResultConflictCopyingFiles
|
|
// but here it's about the base src url being moved/renamed
|
|
// (*m_currentStatSrc) and its dest (m_dest), not about a single file.
|
|
// It also means we already stated the dest, here.
|
|
// On the other hand we haven't stated the src yet (we skipped doing it
|
|
// to save time, since it's not necessary to rename directly!)...
|
|
|
|
Q_ASSERT( m_currentSrcURL == *m_currentStatSrc );
|
|
|
|
// Existing dest?
|
|
if ( ( err == ERR_DIR_ALREADY_EXIST ||
|
|
err == ERR_FILE_ALREADY_EXIST ||
|
|
err == ERR_IDENTICAL_FILES )
|
|
&& isInteractive() )
|
|
{
|
|
if (m_reportTimer)
|
|
m_reportTimer->stop();
|
|
|
|
// Should we skip automatically ?
|
|
if ( m_bAutoSkip ) {
|
|
// Move on to next file
|
|
skipSrc();
|
|
return;
|
|
} else if ( m_bOverwriteAll ) {
|
|
; // nothing to do, stat+copy+del will overwrite
|
|
} else {
|
|
TQString newPath;
|
|
// If src==dest, use "overwrite-itself"
|
|
RenameDlg_Mode mode = (RenameDlg_Mode)
|
|
( ( m_currentSrcURL == dest ) ? M_OVERWRITE_ITSELF : M_OVERWRITE );
|
|
|
|
if ( m_srcList.count() > 1 )
|
|
mode = (RenameDlg_Mode) ( mode | M_MULTI | M_SKIP );
|
|
else
|
|
mode = (RenameDlg_Mode) ( mode | M_SINGLE );
|
|
|
|
// we lack mtime info for both the src (not stated)
|
|
// and the dest (stated but this info wasn't stored)
|
|
// Let's do it for local files, at least
|
|
KIO::filesize_t sizeSrc = (KIO::filesize_t) -1;
|
|
KIO::filesize_t sizeDest = (KIO::filesize_t) -1;
|
|
time_t ctimeSrc = (time_t) -1;
|
|
time_t ctimeDest = (time_t) -1;
|
|
time_t mtimeSrc = (time_t) -1;
|
|
time_t mtimeDest = (time_t) -1;
|
|
|
|
KDE_struct_stat stat_buf;
|
|
if ( m_currentSrcURL.isLocalFile() &&
|
|
KDE_stat(TQFile::encodeName(m_currentSrcURL.path()), &stat_buf) == 0 ) {
|
|
sizeSrc = stat_buf.st_size;
|
|
ctimeSrc = stat_buf.st_ctime;
|
|
mtimeSrc = stat_buf.st_mtime;
|
|
}
|
|
if ( dest.isLocalFile() &&
|
|
KDE_stat(TQFile::encodeName(dest.path()), &stat_buf) == 0 ) {
|
|
sizeDest = stat_buf.st_size;
|
|
ctimeDest = stat_buf.st_ctime;
|
|
mtimeDest = stat_buf.st_mtime;
|
|
}
|
|
|
|
RenameDlg_Result r = Observer::self()->open_RenameDlg(
|
|
this,
|
|
err != ERR_DIR_ALREADY_EXIST ? i18n("File Already Exists") : i18n("Already Exists as Folder"),
|
|
m_currentSrcURL.url(),
|
|
dest.url(),
|
|
mode, newPath,
|
|
sizeSrc, sizeDest,
|
|
ctimeSrc, ctimeDest,
|
|
mtimeSrc, mtimeDest );
|
|
if (m_reportTimer)
|
|
m_reportTimer->start(REPORT_TIMEOUT,false);
|
|
|
|
switch ( r )
|
|
{
|
|
case R_CANCEL:
|
|
{
|
|
m_error = ERR_USER_CANCELED;
|
|
emitResult();
|
|
return;
|
|
}
|
|
case R_RENAME:
|
|
{
|
|
// Set m_dest to the chosen destination
|
|
// This is only for this src url; the next one will revert to d->m_globalDest
|
|
m_dest.setPath( newPath );
|
|
KIO::Job* job = KIO::stat( m_dest, false, 2, false );
|
|
state = STATE_STATING;
|
|
destinationState = DEST_NOT_STATED;
|
|
addSubjob(job);
|
|
return;
|
|
}
|
|
case R_AUTO_SKIP:
|
|
m_bAutoSkip = true;
|
|
// fall through
|
|
case R_SKIP:
|
|
// Move on to next file
|
|
skipSrc();
|
|
return;
|
|
case R_OVERWRITE_ALL:
|
|
m_bOverwriteAll = true;
|
|
break;
|
|
case R_OVERWRITE:
|
|
// Add to overwrite list
|
|
// Note that we add dest, not m_dest.
|
|
// This ensures that when moving several urls into a dir (m_dest),
|
|
// we only overwrite for the current one, not for all.
|
|
// When renaming a single file (m_asMethod), it makes no difference.
|
|
kdDebug(7007) << "adding to overwrite list: " << dest.path() << endl;
|
|
m_overwriteList.append( dest.path() );
|
|
break;
|
|
default:
|
|
//assert( 0 );
|
|
break;
|
|
}
|
|
}
|
|
} else if ( err != KIO::ERR_UNSUPPORTED_ACTION ) {
|
|
kdDebug(7007) << "Couldn't rename " << m_currentSrcURL << " to " << dest << ", aborting" << endl;
|
|
m_error = err;
|
|
m_errorText = errText;
|
|
emitResult();
|
|
return;
|
|
}
|
|
kdDebug(7007) << "Couldn't rename " << m_currentSrcURL << " to " << dest << ", reverting to normal way, starting with stat" << endl;
|
|
//kdDebug(7007) << "KIO::stat on " << m_currentSrcURL << endl;
|
|
KIO::Job* job = KIO::stat( m_currentSrcURL, true, 2, false );
|
|
state = STATE_STATING;
|
|
addSubjob(job);
|
|
m_bOnlyRenames = false;
|
|
}
|
|
else
|
|
{
|
|
//kdDebug(7007) << "Renaming succeeded, move on" << endl;
|
|
emit copyingDone( this, *m_currentStatSrc, dest, true, true );
|
|
statNextSrc();
|
|
}
|
|
}
|
|
|
|
void CopyJob::slotResult( Job *job )
|
|
{
|
|
//kdDebug(7007) << "CopyJob::slotResult() state=" << (int) state << endl;
|
|
// In each case, what we have to do is :
|
|
// 1 - check for errors and treat them
|
|
// 2 - subjobs.remove(job);
|
|
// 3 - decide what to do next
|
|
|
|
switch ( state ) {
|
|
case STATE_STATING: // We were trying to stat a src url or the dest
|
|
slotResultStating( job );
|
|
break;
|
|
case STATE_RENAMING: // We were trying to do a direct renaming, before even stat'ing
|
|
{
|
|
slotResultRenaming( job );
|
|
break;
|
|
}
|
|
case STATE_LISTING: // recursive listing finished
|
|
//kdDebug(7007) << "totalSize: " << (unsigned int) m_totalSize << " files: " << files.count() << " dirs: " << dirs.count() << endl;
|
|
// Was there an error ?
|
|
if (job->error())
|
|
{
|
|
Job::slotResult( job ); // will set the error and emit result(this)
|
|
return;
|
|
}
|
|
|
|
subjobs.remove( job );
|
|
assert ( subjobs.isEmpty() );
|
|
|
|
statNextSrc();
|
|
break;
|
|
case STATE_CREATING_DIRS:
|
|
slotResultCreatingDirs( job );
|
|
break;
|
|
case STATE_CONFLICT_CREATING_DIRS:
|
|
slotResultConflictCreatingDirs( job );
|
|
break;
|
|
case STATE_COPYING_FILES:
|
|
slotResultCopyingFiles( job );
|
|
break;
|
|
case STATE_CONFLICT_COPYING_FILES:
|
|
slotResultConflictCopyingFiles( job );
|
|
break;
|
|
case STATE_DELETING_DIRS:
|
|
slotResultDeletingDirs( job );
|
|
break;
|
|
case STATE_SETTING_DIR_ATTRIBUTES: // TODO KDE4
|
|
assert( 0 );
|
|
//slotResultSettingDirAttributes( job );
|
|
break;
|
|
default:
|
|
assert( 0 );
|
|
}
|
|
}
|
|
|
|
void KIO::CopyJob::setDefaultPermissions( bool b )
|
|
{
|
|
d->m_defaultPermissions = b;
|
|
}
|
|
|
|
// KDE4: remove
|
|
void KIO::CopyJob::setInteractive( bool b )
|
|
{
|
|
Job::setInteractive( b );
|
|
}
|
|
|
|
CopyJob *KIO::copy(const KURL& src, const KURL& dest, bool showProgressInfo )
|
|
{
|
|
//kdDebug(7007) << "KIO::copy src=" << src << " dest=" << dest << endl;
|
|
KURL::List srcList;
|
|
srcList.append( src );
|
|
return new CopyJob( srcList, dest, CopyJob::Copy, false, showProgressInfo );
|
|
}
|
|
|
|
CopyJob *KIO::copyAs(const KURL& src, const KURL& dest, bool showProgressInfo )
|
|
{
|
|
//kdDebug(7007) << "KIO::copyAs src=" << src << " dest=" << dest << endl;
|
|
KURL::List srcList;
|
|
srcList.append( src );
|
|
return new CopyJob( srcList, dest, CopyJob::Copy, true, showProgressInfo );
|
|
}
|
|
|
|
CopyJob *KIO::copy( const KURL::List& src, const KURL& dest, bool showProgressInfo )
|
|
{
|
|
//kdDebug(7007) << src << " " << dest << endl;
|
|
return new CopyJob( src, dest, CopyJob::Copy, false, showProgressInfo );
|
|
}
|
|
|
|
CopyJob *KIO::move(const KURL& src, const KURL& dest, bool showProgressInfo )
|
|
{
|
|
//kdDebug(7007) << src << " " << dest << endl;
|
|
KURL::List srcList;
|
|
srcList.append( src );
|
|
return new CopyJob( srcList, dest, CopyJob::Move, false, showProgressInfo );
|
|
}
|
|
|
|
CopyJob *KIO::moveAs(const KURL& src, const KURL& dest, bool showProgressInfo )
|
|
{
|
|
//kdDebug(7007) << src << " " << dest << endl;
|
|
KURL::List srcList;
|
|
srcList.append( src );
|
|
return new CopyJob( srcList, dest, CopyJob::Move, true, showProgressInfo );
|
|
}
|
|
|
|
CopyJob *KIO::move( const KURL::List& src, const KURL& dest, bool showProgressInfo )
|
|
{
|
|
//kdDebug(7007) << src << " " << dest << endl;
|
|
return new CopyJob( src, dest, CopyJob::Move, false, showProgressInfo );
|
|
}
|
|
|
|
CopyJob *KIO::link(const KURL& src, const KURL& destDir, bool showProgressInfo )
|
|
{
|
|
KURL::List srcList;
|
|
srcList.append( src );
|
|
return new CopyJob( srcList, destDir, CopyJob::Link, false, showProgressInfo );
|
|
}
|
|
|
|
CopyJob *KIO::link(const KURL::List& srcList, const KURL& destDir, bool showProgressInfo )
|
|
{
|
|
return new CopyJob( srcList, destDir, CopyJob::Link, false, showProgressInfo );
|
|
}
|
|
|
|
CopyJob *KIO::linkAs(const KURL& src, const KURL& destDir, bool showProgressInfo )
|
|
{
|
|
KURL::List srcList;
|
|
srcList.append( src );
|
|
return new CopyJob( srcList, destDir, CopyJob::Link, false, showProgressInfo );
|
|
}
|
|
|
|
CopyJob *KIO::trash(const KURL& src, bool showProgressInfo )
|
|
{
|
|
KURL::List srcList;
|
|
srcList.append( src );
|
|
return new CopyJob( srcList, KURL( "trash:/" ), CopyJob::Move, false, showProgressInfo );
|
|
}
|
|
|
|
CopyJob *KIO::trash(const KURL::List& srcList, bool showProgressInfo )
|
|
{
|
|
return new CopyJob( srcList, KURL( "trash:/" ), CopyJob::Move, false, showProgressInfo );
|
|
}
|
|
|
|
//////////
|
|
|
|
DeleteJob::DeleteJob( const KURL::List& src, bool /*shred*/, bool showProgressInfo )
|
|
: Job(showProgressInfo), m_totalSize( 0 ), m_processedSize( 0 ), m_fileProcessedSize( 0 ),
|
|
m_processedFiles( 0 ), m_processedDirs( 0 ), m_totalFilesDirs( 0 ),
|
|
m_srcList(src), m_currentStat(m_srcList.begin()), m_reportTimer(0)
|
|
{
|
|
if ( showProgressInfo ) {
|
|
|
|
connect( this, TQT_SIGNAL( totalFiles( KIO::Job*, unsigned long ) ),
|
|
Observer::self(), TQT_SLOT( slotTotalFiles( KIO::Job*, unsigned long ) ) );
|
|
|
|
connect( this, TQT_SIGNAL( totalDirs( KIO::Job*, unsigned long ) ),
|
|
Observer::self(), TQT_SLOT( slotTotalDirs( KIO::Job*, unsigned long ) ) );
|
|
|
|
// See slotReport
|
|
/*connect( this, TQT_SIGNAL( processedFiles( KIO::Job*, unsigned long ) ),
|
|
m_observer, TQT_SLOT( slotProcessedFiles( KIO::Job*, unsigned long ) ) );
|
|
|
|
connect( this, TQT_SIGNAL( processedDirs( KIO::Job*, unsigned long ) ),
|
|
m_observer, TQT_SLOT( slotProcessedDirs( KIO::Job*, unsigned long ) ) );
|
|
|
|
connect( this, TQT_SIGNAL( deleting( KIO::Job*, const KURL& ) ),
|
|
m_observer, TQT_SLOT( slotDeleting( KIO::Job*, const KURL& ) ) );*/
|
|
|
|
m_reportTimer=new TQTimer(this);
|
|
connect(m_reportTimer,TQT_SIGNAL(timeout()),this,TQT_SLOT(slotReport()));
|
|
//this will update the report dialog with 5 Hz, I think this is fast enough, aleXXX
|
|
m_reportTimer->start(REPORT_TIMEOUT,false);
|
|
}
|
|
|
|
TQTimer::singleShot(0, this, TQT_SLOT(slotStart()));
|
|
}
|
|
|
|
void DeleteJob::slotStart()
|
|
{
|
|
statNextSrc();
|
|
}
|
|
|
|
//this is called often, so calling the functions
|
|
//from Observer here directly might improve the performance a little bit
|
|
//aleXXX
|
|
void DeleteJob::slotReport()
|
|
{
|
|
if (m_progressId==0)
|
|
return;
|
|
|
|
Observer * observer = Observer::self();
|
|
|
|
emit deleting( this, m_currentURL );
|
|
observer->slotDeleting(this,m_currentURL);
|
|
|
|
switch( state ) {
|
|
case STATE_STATING:
|
|
case STATE_LISTING:
|
|
emit totalSize( this, m_totalSize );
|
|
emit totalFiles( this, files.count() );
|
|
emit totalDirs( this, dirs.count() );
|
|
break;
|
|
case STATE_DELETING_DIRS:
|
|
emit processedDirs( this, m_processedDirs );
|
|
observer->slotProcessedDirs(this,m_processedDirs);
|
|
emitPercent( m_processedFiles + m_processedDirs, m_totalFilesDirs );
|
|
break;
|
|
case STATE_DELETING_FILES:
|
|
observer->slotProcessedFiles(this,m_processedFiles);
|
|
emit processedFiles( this, m_processedFiles );
|
|
emitPercent( m_processedFiles, m_totalFilesDirs );
|
|
break;
|
|
}
|
|
}
|
|
|
|
|
|
void DeleteJob::slotEntries(KIO::Job* job, const UDSEntryList& list)
|
|
{
|
|
UDSEntryListConstIterator it = list.begin();
|
|
UDSEntryListConstIterator end = list.end();
|
|
for (; it != end; ++it)
|
|
{
|
|
UDSEntry::ConstIterator it2 = (*it).begin();
|
|
bool bDir = false;
|
|
bool bLink = false;
|
|
TQString displayName;
|
|
KURL url;
|
|
int atomsFound(0);
|
|
for( ; it2 != (*it).end(); it2++ )
|
|
{
|
|
switch ((*it2).m_uds)
|
|
{
|
|
case UDS_FILE_TYPE:
|
|
bDir = S_ISDIR((*it2).m_long);
|
|
atomsFound++;
|
|
break;
|
|
case UDS_NAME:
|
|
displayName = (*it2).m_str;
|
|
atomsFound++;
|
|
break;
|
|
case UDS_URL:
|
|
url = KURL((*it2).m_str);
|
|
atomsFound++;
|
|
break;
|
|
case UDS_LINK_DEST:
|
|
bLink = !(*it2).m_str.isEmpty();
|
|
atomsFound++;
|
|
break;
|
|
case UDS_SIZE:
|
|
m_totalSize += (KIO::filesize_t)((*it2).m_long);
|
|
atomsFound++;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
if (atomsFound==5) break;
|
|
}
|
|
assert(!displayName.isEmpty());
|
|
if (displayName != ".." && displayName != ".")
|
|
{
|
|
if( url.isEmpty() ) {
|
|
url = ((SimpleJob *)job)->url(); // assumed to be a dir
|
|
url.addPath( displayName );
|
|
}
|
|
//kdDebug(7007) << "DeleteJob::slotEntries " << displayName << " (" << url << ")" << endl;
|
|
if ( bLink )
|
|
symlinks.append( url );
|
|
else if ( bDir )
|
|
dirs.append( url );
|
|
else
|
|
files.append( url );
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
void DeleteJob::statNextSrc()
|
|
{
|
|
//kdDebug(7007) << "statNextSrc" << endl;
|
|
if ( m_currentStat != m_srcList.end() )
|
|
{
|
|
m_currentURL = (*m_currentStat);
|
|
|
|
// if the file system doesn't support deleting, we do not even stat
|
|
if (!KProtocolInfo::supportsDeleting(m_currentURL)) {
|
|
TQGuardedPtr<DeleteJob> that = this;
|
|
++m_currentStat;
|
|
if (isInteractive())
|
|
KMessageBox::information( 0, buildErrorString(ERR_CANNOT_DELETE, m_currentURL.prettyURL()));
|
|
if (that)
|
|
statNextSrc();
|
|
return;
|
|
}
|
|
// Stat it
|
|
state = STATE_STATING;
|
|
KIO::SimpleJob * job = KIO::stat( m_currentURL, true, 1, false );
|
|
Scheduler::scheduleJob(job);
|
|
//kdDebug(7007) << "KIO::stat (DeleteJob) " << m_currentURL << endl;
|
|
addSubjob(job);
|
|
//if ( m_progressId ) // Did we get an ID from the observer ?
|
|
// Observer::self()->slotDeleting( this, *it ); // show asap
|
|
} else
|
|
{
|
|
m_totalFilesDirs = files.count()+symlinks.count() + dirs.count();
|
|
slotReport();
|
|
// Now we know which dirs hold the files we're going to delete.
|
|
// To speed things up and prevent double-notification, we disable KDirWatch
|
|
// on those dirs temporarily (using KDirWatch::self, that's the instanced
|
|
// used by e.g. kdirlister).
|
|
for ( TQStringList::Iterator it = m_parentDirs.begin() ; it != m_parentDirs.end() ; ++it )
|
|
KDirWatch::self()->stopDirScan( *it );
|
|
state = STATE_DELETING_FILES;
|
|
deleteNextFile();
|
|
}
|
|
}
|
|
|
|
void DeleteJob::deleteNextFile()
|
|
{
|
|
//kdDebug(7007) << "deleteNextFile" << endl;
|
|
if ( !files.isEmpty() || !symlinks.isEmpty() )
|
|
{
|
|
SimpleJob *job;
|
|
do {
|
|
// Take first file to delete out of list
|
|
KURL::List::Iterator it = files.begin();
|
|
bool isLink = false;
|
|
if ( it == files.end() ) // No more files
|
|
{
|
|
it = symlinks.begin(); // Pick up a symlink to delete
|
|
isLink = true;
|
|
}
|
|
// Normal deletion
|
|
// If local file, try do it directly
|
|
if ( (*it).isLocalFile() && unlink( TQFile::encodeName((*it).path()) ) == 0 ) {
|
|
//kdDebug(7007) << "DeleteJob deleted " << (*it).path() << endl;
|
|
job = 0;
|
|
m_processedFiles++;
|
|
if ( m_processedFiles % 300 == 0 || m_totalFilesDirs < 300) { // update progress info every 300 files
|
|
m_currentURL = *it;
|
|
slotReport();
|
|
}
|
|
} else
|
|
{ // if remote - or if unlink() failed (we'll use the job's error handling in that case)
|
|
job = KIO::file_delete( *it, false /*no GUI*/);
|
|
Scheduler::scheduleJob(job);
|
|
m_currentURL=(*it);
|
|
}
|
|
if ( isLink )
|
|
symlinks.remove(it);
|
|
else
|
|
files.remove(it);
|
|
if ( job ) {
|
|
addSubjob(job);
|
|
return;
|
|
}
|
|
// loop only if direct deletion worked (job=0) and there is something else to delete
|
|
} while (!job && (!files.isEmpty() || !symlinks.isEmpty()));
|
|
}
|
|
state = STATE_DELETING_DIRS;
|
|
deleteNextDir();
|
|
}
|
|
|
|
void DeleteJob::deleteNextDir()
|
|
{
|
|
if ( !dirs.isEmpty() ) // some dirs to delete ?
|
|
{
|
|
do {
|
|
// Take first dir to delete out of list - last ones first !
|
|
KURL::List::Iterator it = dirs.fromLast();
|
|
// If local dir, try to rmdir it directly
|
|
if ( (*it).isLocalFile() && ::rmdir( TQFile::encodeName((*it).path()) ) == 0 ) {
|
|
|
|
m_processedDirs++;
|
|
if ( m_processedDirs % 100 == 0 ) { // update progress info every 100 dirs
|
|
m_currentURL = *it;
|
|
slotReport();
|
|
}
|
|
} else {
|
|
SimpleJob* job;
|
|
if ( KProtocolInfo::canDeleteRecursive( *it ) ) {
|
|
// If the ioslave supports recursive deletion of a directory, then
|
|
// we only need to send a single CMD_DEL command, so we use file_delete :)
|
|
job = KIO::file_delete( *it, false /*no gui*/ );
|
|
} else {
|
|
job = KIO::rmdir( *it );
|
|
}
|
|
Scheduler::scheduleJob(job);
|
|
dirs.remove(it);
|
|
addSubjob( job );
|
|
return;
|
|
}
|
|
dirs.remove(it);
|
|
} while ( !dirs.isEmpty() );
|
|
}
|
|
|
|
// Re-enable watching on the dirs that held the deleted files
|
|
for ( TQStringList::Iterator it = m_parentDirs.begin() ; it != m_parentDirs.end() ; ++it )
|
|
KDirWatch::self()->restartDirScan( *it );
|
|
|
|
// Finished - tell the world
|
|
if ( !m_srcList.isEmpty() )
|
|
{
|
|
KDirNotify_stub allDirNotify("*", "KDirNotify*");
|
|
//kdDebug(7007) << "KDirNotify'ing FilesRemoved " << m_srcList.toStringList() << endl;
|
|
allDirNotify.FilesRemoved( m_srcList );
|
|
}
|
|
if (m_reportTimer!=0)
|
|
m_reportTimer->stop();
|
|
emitResult();
|
|
}
|
|
|
|
void DeleteJob::slotProcessedSize( KIO::Job*, KIO::filesize_t data_size )
|
|
{
|
|
// Note: this is the same implementation as CopyJob::slotProcessedSize but
|
|
// it's different from FileCopyJob::slotProcessedSize - which is why this
|
|
// is not in Job.
|
|
|
|
m_fileProcessedSize = data_size;
|
|
setProcessedSize(m_processedSize + m_fileProcessedSize);
|
|
|
|
//kdDebug(7007) << "DeleteJob::slotProcessedSize " << (unsigned int) (m_processedSize + m_fileProcessedSize) << endl;
|
|
|
|
emit processedSize( this, m_processedSize + m_fileProcessedSize );
|
|
|
|
// calculate percents
|
|
unsigned long ipercent = m_percent;
|
|
|
|
if ( m_totalSize == 0 )
|
|
m_percent = 100;
|
|
else
|
|
m_percent = (unsigned long)(( (float)(m_processedSize + m_fileProcessedSize) / (float)m_totalSize ) * 100.0);
|
|
|
|
if ( m_percent > ipercent )
|
|
{
|
|
emit percent( this, m_percent );
|
|
//kdDebug(7007) << "DeleteJob::slotProcessedSize - percent = " << (unsigned int) m_percent << endl;
|
|
}
|
|
|
|
}
|
|
|
|
void DeleteJob::slotResult( Job *job )
|
|
{
|
|
switch ( state )
|
|
{
|
|
case STATE_STATING:
|
|
{
|
|
// Was there an error while stating ?
|
|
if (job->error() )
|
|
{
|
|
// Probably : doesn't exist
|
|
Job::slotResult( job ); // will set the error and emit result(this)
|
|
return;
|
|
}
|
|
|
|
// Is it a file or a dir ?
|
|
UDSEntry entry = ((StatJob*)job)->statResult();
|
|
bool bDir = false;
|
|
bool bLink = false;
|
|
KIO::filesize_t size = (KIO::filesize_t)-1;
|
|
UDSEntry::ConstIterator it2 = entry.begin();
|
|
int atomsFound(0);
|
|
for( ; it2 != entry.end(); it2++ )
|
|
{
|
|
if ( ((*it2).m_uds) == UDS_FILE_TYPE )
|
|
{
|
|
bDir = S_ISDIR( (mode_t)(*it2).m_long );
|
|
atomsFound++;
|
|
}
|
|
else if ( ((*it2).m_uds) == UDS_LINK_DEST )
|
|
{
|
|
bLink = !((*it2).m_str.isEmpty());
|
|
atomsFound++;
|
|
}
|
|
else if ( ((*it2).m_uds) == UDS_SIZE )
|
|
{
|
|
size = (*it2).m_long;
|
|
atomsFound++;
|
|
}
|
|
if (atomsFound==3) break;
|
|
}
|
|
|
|
KURL url = ((SimpleJob*)job)->url();
|
|
|
|
subjobs.remove( job );
|
|
assert( subjobs.isEmpty() );
|
|
|
|
if (bDir && !bLink)
|
|
{
|
|
// Add toplevel dir in list of dirs
|
|
dirs.append( url );
|
|
if ( url.isLocalFile() && !m_parentDirs.contains( url.path(-1) ) )
|
|
m_parentDirs.append( url.path(-1) );
|
|
|
|
if ( !KProtocolInfo::canDeleteRecursive( url ) ) {
|
|
//kdDebug(7007) << " Target is a directory " << endl;
|
|
// List it
|
|
state = STATE_LISTING;
|
|
ListJob *newjob = listRecursive( url, false );
|
|
newjob->setUnrestricted(true); // No KIOSK restrictions
|
|
Scheduler::scheduleJob(newjob);
|
|
connect(newjob, TQT_SIGNAL(entries( KIO::Job *,
|
|
const KIO::UDSEntryList& )),
|
|
TQT_SLOT( slotEntries( KIO::Job*,
|
|
const KIO::UDSEntryList& )));
|
|
addSubjob(newjob);
|
|
} else {
|
|
++m_currentStat;
|
|
statNextSrc();
|
|
}
|
|
}
|
|
else
|
|
{
|
|
if ( bLink ) {
|
|
//kdDebug(7007) << " Target is a symlink" << endl;
|
|
symlinks.append( url );
|
|
} else {
|
|
//kdDebug(7007) << " Target is a file" << endl;
|
|
files.append( url );
|
|
}
|
|
if ( url.isLocalFile() && !m_parentDirs.contains( url.directory(false) ) )
|
|
m_parentDirs.append( url.directory(false) );
|
|
++m_currentStat;
|
|
statNextSrc();
|
|
}
|
|
}
|
|
break;
|
|
case STATE_LISTING:
|
|
if ( job->error() )
|
|
{
|
|
// Try deleting nonetheless, it may be empty (and non-listable)
|
|
}
|
|
subjobs.remove( job );
|
|
assert( subjobs.isEmpty() );
|
|
++m_currentStat;
|
|
statNextSrc();
|
|
break;
|
|
case STATE_DELETING_FILES:
|
|
if ( job->error() )
|
|
{
|
|
Job::slotResult( job ); // will set the error and emit result(this)
|
|
return;
|
|
}
|
|
subjobs.remove( job );
|
|
assert( subjobs.isEmpty() );
|
|
m_processedFiles++;
|
|
|
|
deleteNextFile();
|
|
break;
|
|
case STATE_DELETING_DIRS:
|
|
if ( job->error() )
|
|
{
|
|
Job::slotResult( job ); // will set the error and emit result(this)
|
|
return;
|
|
}
|
|
subjobs.remove( job );
|
|
assert( subjobs.isEmpty() );
|
|
m_processedDirs++;
|
|
//emit processedDirs( this, m_processedDirs );
|
|
//if (!m_shred)
|
|
//emitPercent( m_processedFiles + m_processedDirs, m_totalFilesDirs );
|
|
|
|
deleteNextDir();
|
|
break;
|
|
default:
|
|
assert(0);
|
|
}
|
|
}
|
|
|
|
DeleteJob *KIO::del( const KURL& src, bool shred, bool showProgressInfo )
|
|
{
|
|
KURL::List srcList;
|
|
srcList.append( src );
|
|
DeleteJob *job = new DeleteJob( srcList, shred, showProgressInfo );
|
|
return job;
|
|
}
|
|
|
|
DeleteJob *KIO::del( const KURL::List& src, bool shred, bool showProgressInfo )
|
|
{
|
|
DeleteJob *job = new DeleteJob( src, shred, showProgressInfo );
|
|
return job;
|
|
}
|
|
|
|
MultiGetJob::MultiGetJob(const KURL& url,
|
|
bool showProgressInfo)
|
|
: TransferJob(url, 0, TQByteArray(), TQByteArray(), showProgressInfo)
|
|
{
|
|
m_waitQueue.setAutoDelete(true);
|
|
m_activeQueue.setAutoDelete(true);
|
|
m_currentEntry = 0;
|
|
}
|
|
|
|
void MultiGetJob::get(long id, const KURL &url, const MetaData &metaData)
|
|
{
|
|
GetRequest *entry = new GetRequest(id, url, metaData);
|
|
entry->metaData["request-id"] = TQString("%1").arg(id);
|
|
m_waitQueue.append(entry);
|
|
}
|
|
|
|
void MultiGetJob::flushQueue(TQPtrList<GetRequest> &queue)
|
|
{
|
|
GetRequest *entry;
|
|
// Use multi-get
|
|
// Scan all jobs in m_waitQueue
|
|
for(entry = m_waitQueue.first(); entry; )
|
|
{
|
|
if ((m_url.protocol() == entry->url.protocol()) &&
|
|
(m_url.host() == entry->url.host()) &&
|
|
(m_url.port() == entry->url.port()) &&
|
|
(m_url.user() == entry->url.user()))
|
|
{
|
|
m_waitQueue.take();
|
|
queue.append(entry);
|
|
entry = m_waitQueue.current();
|
|
}
|
|
else
|
|
{
|
|
entry = m_waitQueue.next();
|
|
}
|
|
}
|
|
// Send number of URLs, (URL, metadata)*
|
|
KIO_ARGS << (TQ_INT32) queue.count();
|
|
for(entry = queue.first(); entry; entry = queue.next())
|
|
{
|
|
stream << entry->url << entry->metaData;
|
|
}
|
|
m_packedArgs = packedArgs;
|
|
m_command = CMD_MULTI_GET;
|
|
m_outgoingMetaData.clear();
|
|
}
|
|
|
|
void MultiGetJob::start(Slave *slave)
|
|
{
|
|
// Add first job from m_waitQueue and add it to m_activeQueue
|
|
GetRequest *entry = m_waitQueue.take(0);
|
|
m_activeQueue.append(entry);
|
|
|
|
m_url = entry->url;
|
|
|
|
if (!entry->url.protocol().startsWith("http"))
|
|
{
|
|
// Use normal get
|
|
KIO_ARGS << entry->url;
|
|
m_packedArgs = packedArgs;
|
|
m_outgoingMetaData = entry->metaData;
|
|
m_command = CMD_GET;
|
|
b_multiGetActive = false;
|
|
}
|
|
else
|
|
{
|
|
flushQueue(m_activeQueue);
|
|
b_multiGetActive = true;
|
|
}
|
|
|
|
TransferJob::start(slave); // Anything else to do??
|
|
}
|
|
|
|
bool MultiGetJob::findCurrentEntry()
|
|
{
|
|
if (b_multiGetActive)
|
|
{
|
|
long id = m_incomingMetaData["request-id"].toLong();
|
|
for(GetRequest *entry = m_activeQueue.first(); entry; entry = m_activeQueue.next())
|
|
{
|
|
if (entry->id == id)
|
|
{
|
|
m_currentEntry = entry;
|
|
return true;
|
|
}
|
|
}
|
|
m_currentEntry = 0;
|
|
return false;
|
|
}
|
|
else
|
|
{
|
|
m_currentEntry = m_activeQueue.first();
|
|
return (m_currentEntry != 0);
|
|
}
|
|
}
|
|
|
|
void MultiGetJob::slotRedirection( const KURL &url)
|
|
{
|
|
if (!findCurrentEntry()) return; // Error
|
|
if (kapp && !kapp->authorizeURLAction("redirect", m_url, url))
|
|
{
|
|
kdWarning(7007) << "MultiGetJob: Redirection from " << m_currentEntry->url << " to " << url << " REJECTED!" << endl;
|
|
return;
|
|
}
|
|
m_redirectionURL = url;
|
|
if (m_currentEntry->url.hasUser() && !url.hasUser() && (m_currentEntry->url.host().lower() == url.host().lower()))
|
|
m_redirectionURL.setUser(m_currentEntry->url.user()); // Preserve user
|
|
get(m_currentEntry->id, m_redirectionURL, m_currentEntry->metaData); // Try again
|
|
}
|
|
|
|
|
|
void MultiGetJob::slotFinished()
|
|
{
|
|
if (!findCurrentEntry()) return;
|
|
if (m_redirectionURL.isEmpty())
|
|
{
|
|
// No redirection, tell the world that we are finished.
|
|
emit result(m_currentEntry->id);
|
|
}
|
|
m_redirectionURL = KURL();
|
|
m_error = 0;
|
|
m_incomingMetaData.clear();
|
|
m_activeQueue.removeRef(m_currentEntry);
|
|
if (m_activeQueue.count() == 0)
|
|
{
|
|
if (m_waitQueue.count() == 0)
|
|
{
|
|
// All done
|
|
TransferJob::slotFinished();
|
|
}
|
|
else
|
|
{
|
|
// return slave to pool
|
|
// fetch new slave for first entry in m_waitQueue and call start
|
|
// again.
|
|
GetRequest *entry = m_waitQueue.tqat(0);
|
|
m_url = entry->url;
|
|
slaveDone();
|
|
Scheduler::doJob(this);
|
|
}
|
|
}
|
|
}
|
|
|
|
void MultiGetJob::slotData( const TQByteArray &_data)
|
|
{
|
|
if(!m_currentEntry) return;// Error, unknown request!
|
|
if(m_redirectionURL.isEmpty() || !m_redirectionURL.isValid() || m_error)
|
|
emit data(m_currentEntry->id, _data);
|
|
}
|
|
|
|
void MultiGetJob::slotMimetype( const TQString &_mimetype )
|
|
{
|
|
if (b_multiGetActive)
|
|
{
|
|
TQPtrList<GetRequest> newQueue;
|
|
flushQueue(newQueue);
|
|
if (!newQueue.isEmpty())
|
|
{
|
|
while(!newQueue.isEmpty())
|
|
m_activeQueue.append(newQueue.take(0));
|
|
m_slave->send( m_command, m_packedArgs );
|
|
}
|
|
}
|
|
if (!findCurrentEntry()) return; // Error, unknown request!
|
|
emit mimetype(m_currentEntry->id, _mimetype);
|
|
}
|
|
|
|
MultiGetJob *KIO::multi_get(long id, const KURL &url, const MetaData &metaData)
|
|
{
|
|
MultiGetJob * job = new MultiGetJob( url, false );
|
|
job->get(id, url, metaData);
|
|
return job;
|
|
}
|
|
|
|
|
|
#ifdef CACHE_INFO
|
|
CacheInfo::CacheInfo(const KURL &url)
|
|
{
|
|
m_url = url;
|
|
}
|
|
|
|
TQString CacheInfo::cachedFileName()
|
|
{
|
|
const TQChar separator = '_';
|
|
|
|
TQString CEF = m_url.path();
|
|
|
|
int p = CEF.find('/');
|
|
|
|
while(p != -1)
|
|
{
|
|
CEF[p] = separator;
|
|
p = CEF.find('/', p);
|
|
}
|
|
|
|
TQString host = m_url.host().lower();
|
|
CEF = host + CEF + '_';
|
|
|
|
TQString dir = KProtocolManager::cacheDir();
|
|
if (dir[dir.length()-1] != '/')
|
|
dir += "/";
|
|
|
|
int l = m_url.host().length();
|
|
for(int i = 0; i < l; i++)
|
|
{
|
|
if (host[i].isLetter() && (host[i] != 'w'))
|
|
{
|
|
dir += host[i];
|
|
break;
|
|
}
|
|
}
|
|
if (dir[dir.length()-1] == '/')
|
|
dir += "0";
|
|
|
|
unsigned long hash = 0x00000000;
|
|
TQCString u = m_url.url().latin1();
|
|
for(int i = u.length(); i--;)
|
|
{
|
|
hash = (hash * 12211 + u[i]) % 2147483563;
|
|
}
|
|
|
|
TQString hashString;
|
|
hashString.sprintf("%08lx", hash);
|
|
|
|
CEF = CEF + hashString;
|
|
|
|
CEF = dir + "/" + CEF;
|
|
|
|
return CEF;
|
|
}
|
|
|
|
TQFile *CacheInfo::cachedFile()
|
|
{
|
|
#ifdef Q_WS_WIN
|
|
const char *mode = (readWrite ? "rb+" : "rb");
|
|
#else
|
|
const char *mode = (readWrite ? "r+" : "r");
|
|
#endif
|
|
|
|
FILE *fs = fopen(TQFile::encodeName(CEF), mode); // Open for reading and writing
|
|
if (!fs)
|
|
return 0;
|
|
|
|
char buffer[401];
|
|
bool ok = true;
|
|
|
|
// CacheRevision
|
|
if (ok && (!fgets(buffer, 400, fs)))
|
|
ok = false;
|
|
if (ok && (strcmp(buffer, CACHE_REVISION) != 0))
|
|
ok = false;
|
|
|
|
time_t date;
|
|
time_t tqcurrentDate = time(0);
|
|
|
|
// URL
|
|
if (ok && (!fgets(buffer, 400, fs)))
|
|
ok = false;
|
|
if (ok)
|
|
{
|
|
int l = strlen(buffer);
|
|
if (l>0)
|
|
buffer[l-1] = 0; // Strip newline
|
|
if (m_.url.url() != buffer)
|
|
{
|
|
ok = false; // Hash collision
|
|
}
|
|
}
|
|
|
|
// Creation Date
|
|
if (ok && (!fgets(buffer, 400, fs)))
|
|
ok = false;
|
|
if (ok)
|
|
{
|
|
date = (time_t) strtoul(buffer, 0, 10);
|
|
if (m_maxCacheAge && (difftime(tqcurrentDate, date) > m_maxCacheAge))
|
|
{
|
|
m_bMustRevalidate = true;
|
|
m_expireDate = tqcurrentDate;
|
|
}
|
|
}
|
|
|
|
// Expiration Date
|
|
m_cacheExpireDateOffset = ftell(fs);
|
|
if (ok && (!fgets(buffer, 400, fs)))
|
|
ok = false;
|
|
if (ok)
|
|
{
|
|
if (m_request.cache == CC_Verify)
|
|
{
|
|
date = (time_t) strtoul(buffer, 0, 10);
|
|
// After the expire date we need to revalidate.
|
|
if (!date || difftime(tqcurrentDate, date) >= 0)
|
|
m_bMustRevalidate = true;
|
|
m_expireDate = date;
|
|
}
|
|
}
|
|
|
|
// ETag
|
|
if (ok && (!fgets(buffer, 400, fs)))
|
|
ok = false;
|
|
if (ok)
|
|
{
|
|
m_etag = TQString(buffer).stripWhiteSpace();
|
|
}
|
|
|
|
// Last-Modified
|
|
if (ok && (!fgets(buffer, 400, fs)))
|
|
ok = false;
|
|
if (ok)
|
|
{
|
|
m_lastModified = TQString(buffer).stripWhiteSpace();
|
|
}
|
|
|
|
fclose(fs);
|
|
|
|
if (ok)
|
|
return fs;
|
|
|
|
unlink( TQFile::encodeName(CEF) );
|
|
return 0;
|
|
|
|
}
|
|
|
|
void CacheInfo::flush()
|
|
{
|
|
cachedFile().remove();
|
|
}
|
|
|
|
void CacheInfo::touch()
|
|
{
|
|
|
|
}
|
|
void CacheInfo::setExpireDate(int);
|
|
void CacheInfo::setExpireTimeout(int);
|
|
|
|
|
|
int CacheInfo::creationDate();
|
|
int CacheInfo::expireDate();
|
|
int CacheInfo::expireTimeout();
|
|
#endif
|
|
|
|
void Job::virtual_hook( int, void* )
|
|
{ /*BASE::virtual_hook( id, data );*/ }
|
|
|
|
void SimpleJob::virtual_hook( int id, void* data )
|
|
{ KIO::Job::virtual_hook( id, data ); }
|
|
|
|
void MkdirJob::virtual_hook( int id, void* data )
|
|
{ SimpleJob::virtual_hook( id, data ); }
|
|
|
|
void StatJob::virtual_hook( int id, void* data )
|
|
{ SimpleJob::virtual_hook( id, data ); }
|
|
|
|
void TransferJob::virtual_hook( int id, void* data )
|
|
{ SimpleJob::virtual_hook( id, data ); }
|
|
|
|
void MultiGetJob::virtual_hook( int id, void* data )
|
|
{ TransferJob::virtual_hook( id, data ); }
|
|
|
|
void MimetypeJob::virtual_hook( int id, void* data )
|
|
{ TransferJob::virtual_hook( id, data ); }
|
|
|
|
void FileCopyJob::virtual_hook( int id, void* data )
|
|
{ Job::virtual_hook( id, data ); }
|
|
|
|
void ListJob::virtual_hook( int id, void* data )
|
|
{ SimpleJob::virtual_hook( id, data ); }
|
|
|
|
void CopyJob::virtual_hook( int id, void* data )
|
|
{ Job::virtual_hook( id, data ); }
|
|
|
|
void DeleteJob::virtual_hook( int id, void* data )
|
|
{ Job::virtual_hook( id, data ); }
|
|
|
|
|
|
#include "jobclasses.moc"
|