# # Copyright (C) 2006 Alon Zakai ('Kripken') # Copyright (C) 2006 Zach Tibbitts # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2, or (at your option) # any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # # Deluge Library, a.k.a. pytorrent: # # Deluge is a client. pytorrent is a Python library for torrenting, that includes # pytorrent.py, which is Python code, and pytorrent_core, which is also a Python # module, but written in C++, and includes the libtorrent torrent library. Only # pytorrent should be visible, and only it should be imported, in the client. # import pytorrent_core import os, shutil import pickle import time # Constants TORRENTS_SUBDIR = "torrentfiles" STATE_FILENAME = "persistent.state" PREFS_FILENAME = "prefs.state" DHT_FILENAME = "dht.state" TORRENT_STATE_EXPIRATION = 1 # seconds, like the output of time.time() DEFAULT_PREFS = { # "max_half_open" : -1, "max_uploads" : 2, # a.k.a. upload slots "listen_on" : [6881,9999], "max_connections" : 80, "use_DHT" : True, "max_active_torrents" : -1, "auto_seed_ratio" : -1, "max_download_rate" : -1, "max_upload_rate" : -1 } # Exception class PyTorrentError(Exception): def __init__(self, value): self.value = value def __str__(self): return repr(self.value) # Information for a single torrent class torrent: def __init__(self, filename, save_dir, compact): self.filename = filename self.save_dir = save_dir self.compact = compact self.user_paused = False # start out unpaused self.uploaded_memory = 0 self.filter_out = [] self.delete_me = False # set this to true, to delete it on next sync # The persistent state of the torrent system. Everything in this will be pickled class persistent_state: def __init__(self): # Torrents self.torrents = [] # Prepare queue (queue is pickled, just like everything else) self.queue = [] # queue[x] is the unique_ID of the x-th queue position. Simple. # The manager for the torrent system class manager: def __init__(self, client_ID, version, user_agent, base_dir): self.base_dir = base_dir # Ensure directories exist if not TORRENTS_SUBDIR in os.listdir(self.base_dir): os.mkdir(self.base_dir + "/" + TORRENTS_SUBDIR) # Start up the core assert(len(version) == 4) pytorrent_core.init(client_ID, int(version[0]), int(version[1]), int(version[2]), int(version[3]), user_agent) self.constants = pytorrent_core.constants() # Unique IDs are NOT in the state, since they are temporary for each session self.unique_IDs = {} # unique_ID -> a torrent object # Saved torrent states. We do not poll the core in a costly manner, necessarily self.saved_torrent_states = {} # unique_ID -> torrent_state self.saved_torrent_states_timestamp = {} # time of creation # Unpickle the preferences, or create a new one try: pkl_file = open(self.base_dir + "/" + PREFS_FILENAME, 'rb') self.prefs = pickle.load(pkl_file) pkl_file.close() except IOError: self.prefs = DEFAULT_PREFS # Apply preferences. Note that this is before any torrents are added self.apply_prefs() # Apply DHT, if needed. Note that this is before any torrents are added if self.get_pref('use_DHT'): pytorrent_core.start_DHT(self.base_dir + "/" + DHT_FILENAME) # Unpickle the state, or create a new one try: pkl_file = open(self.base_dir + "/" + STATE_FILENAME, 'rb') self.state = pickle.load(pkl_file) pkl_file.close() # Sync with the core: tell core about torrents, and get unique_IDs self.sync() except IOError: self.state = persistent_state() def quit(self): # Pickle the prefs output = open(self.base_dir + "/" + PREFS_FILENAME, 'wb') pickle.dump(self.prefs, output) output.close() # Pickle the state output = open(self.base_dir + "/" + STATE_FILENAME, 'wb') pickle.dump(self.state, output) output.close() # Save fastresume data self.save_fastresume_data() # Stop DHT, if needed if self.get_pref('use_DHT'): pytorrent_core.stop_DHT(self.base_dir + "/" + DHT_FILENAME) # Shutdown torrent core pytorrent_core.quit() def get_pref(self, key): # If we have a value, return, else fallback on default_prefs, else raise an error # the fallback is useful if the source has newer prefs than the existing pref state, # which was created by an old version of the source if key in self.prefs.keys(): return self.prefs[key] elif key in DEFAULT_PREFS: self.prefs[key] = DEFAULT_PREFS[key] return self.prefs[key] else: raise PyTorrentError("Asked for a pref that doesn't exist: " + key) def set_pref(self, key, value): # Make sure this is a valid key if key not in DEFAULT_PREFS.keys(): raise PyTorrentError("Asked to change a pref that isn't valid: " + key) self.prefs[key] = value def apply_prefs(self): print "Applying preferences" pytorrent_core.set_download_rate_limit(self.get_pref('max_download_rate')*1024) pytorrent_core.set_upload_rate_limit(self.get_pref('max_upload_rate')*1024) pytorrent_core.set_listen_on(self.get_pref('listen_on')[0], self.get_pref('listen_on')[1]) pytorrent_core.set_max_connections(self.get_pref('max_connections')) pytorrent_core.set_max_uploads(self.get_pref('max_uploads')) def add_torrent(self, filename, save_dir, compact): self.add_torrent_ns(filename, save_dir, compact) return self.sync() # Syncing will create a new torrent in the core, and return it's ID def remove_torrent(self, unique_ID, data_also): # Save some data before we remove the torrent, needed later in this func temp = self.unique_IDs[unique_ID] temp_fileinfo = pytorrent_core.get_fileinfo(unique_ID) self.remove_torrent_ns(unique_ID) self.sync() # Remove .torrent and .fastresume os.remove(temp.filename) try: # Must be after removal of the torrent, because that saves a new .fastresume os.remove(temp.filename + ".fastresume") except OSError: pass # Perhaps there never was one to begin with # Remove data, if asked to do so if data_also: # Must be done AFTER the torrent is removed # Note: can this be to the trash? for filedata in temp_fileinfo: filename = filedata['path'] try: os.remove(temp.save_dir + "/" + filename) except OSError: pass # No file just means it wasn't downloaded, we can continue # A separate function, because people may want to call it from time to time def save_fastresume_data(self): for unique_ID in self.unique_IDs: pytorrent_core.save_fastresume(unique_ID, self.unique_IDs[unique_ID].filename) def get_state(self): ret = pytorrent_core.get_session_info() ret['is_listening'] = pytorrent_core.is_listening() ret['port'] = pytorrent_core.listening_port() if self.get_pref('use_DHT'): ret['DHT_nodes'] = pytorrent_core.get_DHT_info() return ret # Efficient: use a saved state, if it hasn't expired yet def get_torrent_state(self, unique_ID, efficiently = False): if efficiently: try: if time.time() < self.saved_torrent_states_timestamp[unique_ID] + \ TORRENT_STATE_EXPIRATION: return self.saved_torrent_states[unique_ID] except KeyError: pass self.saved_torrent_states_timestamp[unique_ID] = time.time() self.saved_torrent_states[unique_ID] = pytorrent_core.get_state(unique_ID) return self.saved_torrent_states[unique_ID] def mark_state_dirty(self, unique_ID): del self.saved_torrent_states[unique_ID] del self.saved_torrent_states_timestamp[unique_ID] def queue_up(self, unique_ID): curr_index = self.get_queue_index(unique_ID) if curr_index > 0: temp = self.state.queue[curr_index - 1] self.state.queue[curr_index - 1] = unique_ID self.state.queue[curr_index] = temp def queue_down(self, unique_ID): curr_index = self.get_queue_index(unique_ID) if curr_index < (len(self.state.queue) - 1): temp = self.state.queue[curr_index + 1] self.state.queue[curr_index + 1] = unique_ID self.state.queue[curr_index] = temp def queue_bottom(self, unique_ID): curr_index = self.get_queue_index(unique_ID) if curr_index < (len(self.state.queue) - 1): self.state.queue.remove(curr_index) self.state.queue.append(unique_ID) def clear_completed(self): for unique_ID in self.unique_IDs: torrent_state = self.get_torrent_state(unique_ID, True) if torrent_state['progress'] == 100.0: self.remove_torrent_ns(unique_ID) self.sync() def set_user_pause(self, unique_ID, new_value): self.unique_IDs[unique_ID].user_paused = new_value self.apply_queue() def is_user_paused(self, unique_ID): return self.unique_IDs[unique_ID].user_paused def get_unique_IDs(self): return self.unique_IDs.keys() # Enforce the queue: pause/unpause as needed, based on queue and user_pausing # This should be called after changes to relevant parameters (user_pausing, or # altering max_active_torrents), or just from time to time # ___ALL queuing code should be in this function, and ONLY here___ def apply_queue(self): # Handle autoseeding - downqueue as needed if self.auto_seed_ratio != -1: for unique_ID in self.unique_IDs: if self.get_torrent_state(unique_ID, True)['is_seed']: torrent_state = self.get_torrent_state(unique_ID, True) ratio = self.calc_ratio(unique_ID, torrent_state) if ratio >= self.auto_seed_ratio: self.queue_bottom(unique_ID) # Pause and resume torrents for index in range(len(self.state.queue)): unique_ID = self.state.queue[index] if (index < self.state.max_active_torrents or self.state_max_active_torrents == -1) \ and self.get_torrent_state(unique_ID, True)['is_paused'] \ and not self.is_user_paused(unique_ID): pytorrent_core.resume(unique_ID) elif not self.get_torrent_state(unique_ID, True)['is_paused'] and \ (index >= self.state.max_active_torrents or self.is_user_paused(unique_ID)): pytorrent_core.pause(unique_ID) def calc_ratio(self, unique_ID, torrent_state): up = float(torrent_state['total_upload'] + self.unique_IDs[unique_ID].uploaded_memory) down = float(torrent_state["total_done"]) try: ret = float(up/down) except: ret = -1 return ret def get_num_torrents(self): return pytorrent_core.get_num_torrents() def handle_events(self): event = pytorrent_core.pop_event() while event is not None: print "EVENT: ", event if event['event_type'] is self.constants['EVENT_FINISHED']: # If we are autoseeding, then we need to apply the queue if self.auto_seed_ratio == -1: self.mark_state_dirty(event['unique_ID']) # So the queuing will be to new data self.apply_queue() elif event['event_type'] is self.constants['EVENT_PEER_ERROR']: # self.parent.addMessage(_("Peer Error") + ": " + str(event), "I") # Debug only! pass elif event['event_type'] is self.constants['EVENT_INVALID_REQUEST']: print 'self.parent.addMessage(_("Invalid request") + ": " + str(event), "W") # Maybe "I"?' elif event['event_type'] is self.constants['EVENT_FILE_ERROR']: # dc.debugmsg("File error! " + str(event)) print 'self.parent.addMessage(_("File error") + "! " + str(event), "F")' elif event['event_type'] is self.constants['EVENT_HASH_FAILED_ERROR']: print 'self.parent.addMessage(_("Hash failed") + ": " + str(event), "I")' elif event['event_type'] is self.constants['EVENT_PEER_BAN_ERROR']: print 'self.parent.addMessage(_("Peer banned") + ": " + str(event), "I")' elif event['event_type'] is self.constants['EVENT_FASTRESUME_REJECTED_ERROR']: # dc.debugmsg("Fastresume rejected: " + str(event)) print 'self.parent.addMessage(_("Fastresume rejected") + ": " + str(event), "W")' elif event['event_type'] is self.constants['EVENT_TRACKER']: print event['tracker_status'], event['message'] elif event['event_type'] is self.constants['EVENT_OTHER']: print 'self.parent.addMessage(_("Event") + ": " + str(event), "W")' else: # dc.debugmsg("Internal error, undefined event type") # dc.debugmsg("No such event error. Raw data: " + str(event)) print 'self.parent.addMessage(_("Event") + ": " + str(event), "C")' event = pytorrent_core.pop_event() #################### # Internal functions #################### # Non-syncing functions. Used when we loop over such events, and sync manually at the end def add_torrent_ns(self, filename, save_dir, compact): # Cache torrent file (temp, torrent_file_short) = os.path.split(filename) time.sleep(0.01) # Ensure we use a unique time for the new filename new_name = str(time.time()) + ".torrent" full_new_name = self.base_dir + "/" + TORRENTS_SUBDIR + "/" + new_name if new_name in os.listdir(self.base_dir + "/" + TORRENTS_SUBDIR): raise PyTorrentError("Could not cache torrent file locally, failed: " + new_name) shutil.copy(filename, full_new_name) # Create torrent object new_torrent = torrent(full_new_name, save_dir, compact) self.state.torrents.append(new_torrent) def remove_torrent_ns(self, unique_ID): self.unique_IDs[unique_ID].delete_me = True # Sync the state.torrents and unique_IDs lists with the core # ___ALL syncing code with the core is here, and ONLY here___ # Also all self-syncing is done here (various lists) def sync(self): ret = None # We return new added unique ID(s), or None # Add torrents to core and unique_IDs torrents_with_unique_ID = self.unique_IDs.values() for torrent in self.state.torrents: if torrent not in torrents_with_unique_ID: # print "Adding torrent to core:", torrent.filename, torrent.save_dir, torrent.compact unique_ID = pytorrent_core.add_torrent(torrent.filename, torrent.save_dir, torrent.compact) # print "Got unique ID:", unique_ID ret = unique_ID self.unique_IDs[unique_ID] = torrent # Remove torrents from core, unique_IDs and queue to_delete = [] for torrent in self.state.torrents: if torrent.delete_me: pytorrent_core.remove_torrent(torrent.unique_ID, torrent.filename) to_delete.append(torrent.unique_ID) for unique_ID in to_delete: self.state.torrents.remove(self.unique_IDs[unique_ID]) self.state.queue.remove(self.unique_IDs[unique_ID]) del self.unique_IDs[unique_ID] # Add torrents to queue - at the end, of course for unique_ID in self.unique_IDs: if unique_ID not in self.state.queue: self.state.queue.append(unique_ID) assert(len(self.unique_IDs) == len(self.state.torrents)) assert(len(self.unique_IDs) == len(self.state.queue)) assert(len(self.unique_IDs) == pytorrent_core.get_num_torrents()) return ret def get_queue_index(self, unique_ID): return self.state.queue.index(unique_ID)