2010-05-15 08:38:45 -07:00
|
|
|
class Item < ActiveRecord::Base
|
2010-05-19 16:17:33 -07:00
|
|
|
SwfAssetType = 'object'
|
2010-05-16 17:45:30 -07:00
|
|
|
|
2010-11-06 15:08:42 -07:00
|
|
|
has_one :contribution, :as => :contributed
|
2010-10-07 07:46:23 -07:00
|
|
|
has_many :parent_swf_asset_relationships, :foreign_key => 'parent_id',
|
|
|
|
:conditions => {:swf_asset_type => SwfAssetType}
|
2010-11-25 18:33:34 -08:00
|
|
|
has_many :swf_assets, :through => :parent_swf_asset_relationships, :source => :object_asset,
|
|
|
|
:conditions => {:type => SwfAssetType}
|
2010-10-07 07:46:23 -07:00
|
|
|
|
2010-10-09 09:22:40 -07:00
|
|
|
attr_writer :current_body_id
|
|
|
|
|
2010-09-08 19:49:39 -07:00
|
|
|
NCRarities = [0, 500]
|
2010-10-03 17:14:48 -07:00
|
|
|
PaintbrushSetDescription = 'This item is part of a deluxe paint brush set!'
|
2010-09-08 19:49:39 -07:00
|
|
|
|
2010-05-15 08:38:45 -07:00
|
|
|
set_table_name 'objects' # Neo & PHP Impress call them objects, but the class name is a conflict (duh!)
|
|
|
|
set_inheritance_column 'inheritance_type' # PHP Impress used "type" to describe category
|
|
|
|
|
2010-05-15 11:14:17 -07:00
|
|
|
cattr_reader :per_page
|
|
|
|
@@per_page = 30
|
|
|
|
|
2010-05-15 11:01:15 -07:00
|
|
|
scope :alphabetize, order('name ASC')
|
|
|
|
|
2010-11-27 15:41:06 -08:00
|
|
|
scope :join_swf_assets, joins("INNER JOIN #{ParentSwfAssetRelationship.table_name} psa ON psa.swf_asset_type = 'object' AND psa.parent_id = objects.id").
|
|
|
|
joins("INNER JOIN #{SwfAsset.table_name} swf_assets ON swf_assets.id = psa.swf_asset_id").
|
2010-06-22 09:42:25 -07:00
|
|
|
group('objects.id')
|
2010-06-09 21:51:53 -07:00
|
|
|
|
2010-11-27 15:41:06 -08:00
|
|
|
scope :without_swf_assets, joins(
|
|
|
|
"LEFT JOIN #{ParentSwfAssetRelationship.table_name} psa ON psa.swf_asset_type = 'object' AND psa.parent_id = #{table_name}.id " +
|
|
|
|
"LEFT JOIN #{SwfAsset.table_name} sa ON sa.type = 'object' AND sa.id = psa.swf_asset_id"
|
|
|
|
).where('sa.id IS NULL')
|
|
|
|
|
|
|
|
scope :spidered_longest_ago, order(["(#{Item.arel_table[:last_spidered].eq(nil).to_sql}) DESC", arel_table[:last_spidered].desc])
|
|
|
|
|
2010-12-05 13:06:48 -08:00
|
|
|
scope :sold_in_mall, where(:sold_in_mall => true)
|
|
|
|
scope :not_sold_in_mall, where(:sold_in_mall => false)
|
2010-11-27 15:41:06 -08:00
|
|
|
|
2010-05-14 15:41:40 -07:00
|
|
|
# Not defining validations, since this app is currently read-only
|
2010-05-15 08:38:45 -07:00
|
|
|
|
2010-09-08 19:49:39 -07:00
|
|
|
def nc?
|
|
|
|
NCRarities.include?(rarity_index)
|
|
|
|
end
|
|
|
|
|
|
|
|
def restricted_zones
|
|
|
|
unless @restricted_zones
|
|
|
|
@restricted_zones = []
|
|
|
|
zones_restrict.split(//).each_with_index do |switch, id|
|
|
|
|
@restricted_zones << Zone.find(id.to_i + 1) if switch == '1'
|
|
|
|
end
|
|
|
|
end
|
|
|
|
@restricted_zones
|
|
|
|
end
|
|
|
|
|
|
|
|
def occupied_zones
|
|
|
|
all_body_ids = []
|
|
|
|
zone_body_ids = {}
|
|
|
|
selected_assets = swf_assets.select('body_id, zone_id').each do |swf_asset|
|
|
|
|
zone_body_ids[swf_asset.zone_id] ||= []
|
|
|
|
body_ids = zone_body_ids[swf_asset.zone_id]
|
|
|
|
body_ids << swf_asset.body_id unless body_ids.include?(swf_asset.body_id)
|
|
|
|
all_body_ids << swf_asset.body_id unless all_body_ids.include?(swf_asset.body_id)
|
|
|
|
end
|
|
|
|
zones = []
|
|
|
|
total_body_ids = all_body_ids.size
|
|
|
|
zone_body_ids.each do |zone_id, body_ids|
|
|
|
|
zone = Zone.find(zone_id)
|
|
|
|
zone.sometimes = true if body_ids.size < total_body_ids
|
|
|
|
zones << zone
|
|
|
|
end
|
|
|
|
zones
|
|
|
|
end
|
|
|
|
|
2010-05-15 08:38:45 -07:00
|
|
|
def species_support_ids
|
2010-11-27 15:41:06 -08:00
|
|
|
@species_support_ids_array ||= read_attribute('species_support_ids').split(',').map(&:to_i) rescue nil
|
2010-05-15 08:38:45 -07:00
|
|
|
end
|
|
|
|
|
|
|
|
def species_support_ids=(replacement)
|
2010-05-16 12:01:38 -07:00
|
|
|
@species_support_ids_array = nil
|
2010-05-15 08:38:45 -07:00
|
|
|
replacement = replacement.join(',') if replacement.is_a?(Array)
|
|
|
|
write_attribute('species_support_ids', replacement)
|
|
|
|
end
|
|
|
|
|
2010-05-16 13:15:21 -07:00
|
|
|
def supported_species
|
2010-11-27 15:41:06 -08:00
|
|
|
@supported_species ||= species_support_ids.blank? ? Species.all : species_support_ids.sort.map { |id| Species.find(id) }
|
2010-05-16 13:15:21 -07:00
|
|
|
end
|
|
|
|
|
2010-05-15 08:38:45 -07:00
|
|
|
def self.search(query)
|
2010-11-16 20:32:53 -08:00
|
|
|
raise SearchError, "Please provide a search query" unless query
|
2010-05-15 11:56:21 -07:00
|
|
|
query = query.strip
|
2010-11-16 20:32:53 -08:00
|
|
|
raise SearchError, "Search queries should be at least 3 characters" if query.length < 3
|
2010-05-15 08:38:45 -07:00
|
|
|
query_conditions = [Condition.new]
|
|
|
|
in_phrase = false
|
|
|
|
query.each_char do |c|
|
|
|
|
if c == ' ' && !in_phrase
|
|
|
|
query_conditions << Condition.new
|
|
|
|
elsif c == '"'
|
|
|
|
in_phrase = !in_phrase
|
|
|
|
elsif c == ':' && !in_phrase
|
2010-06-22 09:42:25 -07:00
|
|
|
query_conditions.last.to_filter!
|
2010-05-15 09:43:54 -07:00
|
|
|
elsif c == '-' && !in_phrase && query_conditions.last.empty?
|
|
|
|
query_conditions.last.negate!
|
2010-05-15 08:38:45 -07:00
|
|
|
else
|
|
|
|
query_conditions.last << c
|
|
|
|
end
|
|
|
|
end
|
2010-11-18 20:21:03 -08:00
|
|
|
limited_filters_used = []
|
2010-06-22 09:42:25 -07:00
|
|
|
query_conditions.inject(self.scoped) do |scope, condition|
|
2010-11-18 20:21:03 -08:00
|
|
|
if condition.filter? && LimitedSearchFilters.include?(condition.filter)
|
|
|
|
if limited_filters_used.include?(condition.filter)
|
|
|
|
raise SearchError, "The #{condition.filter} filter is complex; please only use one per search. Thanks!"
|
|
|
|
else
|
|
|
|
limited_filters_used << condition.filter
|
|
|
|
end
|
|
|
|
end
|
2010-05-15 08:38:45 -07:00
|
|
|
condition.narrow(scope)
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2010-06-22 10:00:55 -07:00
|
|
|
def as_json(options = {})
|
|
|
|
{
|
|
|
|
:description => description,
|
|
|
|
:id => id,
|
|
|
|
:name => name,
|
|
|
|
:thumbnail_url => thumbnail_url,
|
2010-09-08 17:07:09 -07:00
|
|
|
:zones_restrict => zones_restrict,
|
|
|
|
:rarity_index => rarity_index
|
2010-06-22 10:00:55 -07:00
|
|
|
}
|
|
|
|
end
|
|
|
|
|
2010-10-10 11:43:01 -07:00
|
|
|
before_create do
|
2010-11-25 08:10:21 -08:00
|
|
|
self.sold_in_mall ||= false
|
2010-10-09 07:53:58 -07:00
|
|
|
true
|
2010-10-07 07:46:23 -07:00
|
|
|
end
|
|
|
|
|
2010-11-06 15:08:42 -07:00
|
|
|
def handle_assets!
|
2010-10-09 09:22:40 -07:00
|
|
|
if @parent_swf_asset_relationships_to_update && @current_body_id
|
|
|
|
new_swf_asset_ids = @parent_swf_asset_relationships_to_update.map(&:swf_asset_id)
|
|
|
|
rels = ParentSwfAssetRelationship.arel_table
|
|
|
|
swf_assets = SwfAsset.arel_table
|
|
|
|
ids_to_delete = self.parent_swf_asset_relationships.
|
|
|
|
select(:id).
|
2010-10-10 19:18:42 -07:00
|
|
|
joins(:object_asset).
|
2010-11-16 14:26:06 -08:00
|
|
|
where(rels[:swf_asset_id].not_in(new_swf_asset_ids)).
|
2010-10-09 09:22:40 -07:00
|
|
|
where(swf_assets[:body_id].in([@current_body_id, 0])).
|
|
|
|
map(&:id)
|
|
|
|
unless ids_to_delete.empty?
|
|
|
|
ParentSwfAssetRelationship.
|
|
|
|
where(rels[:parent_id].eq(self.id)).
|
|
|
|
where(rels[:swf_asset_type].eq(SwfAssetType)).
|
|
|
|
where(rels[:swf_asset_id].in(ids_to_delete)).
|
|
|
|
delete_all
|
|
|
|
end
|
2010-11-06 15:08:42 -07:00
|
|
|
self.parent_swf_asset_relationships += @parent_swf_asset_relationships_to_update
|
2010-10-09 09:22:40 -07:00
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2010-10-07 07:46:23 -07:00
|
|
|
def origin_registry_info=(info)
|
|
|
|
# bear in mind that numbers from registries are floats
|
|
|
|
self.species_support_ids = info[:species_support].map(&:to_i)
|
|
|
|
attribute_names.each do |attribute|
|
|
|
|
value = info[attribute.to_sym]
|
|
|
|
if value
|
|
|
|
value = value.to_i if value.is_a? Float
|
|
|
|
self[attribute] = value
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2010-11-06 15:08:42 -07:00
|
|
|
def pending_swf_assets
|
|
|
|
@parent_swf_asset_relationships_to_update.inject([]) do |all_swf_assets, relationship|
|
|
|
|
all_swf_assets << relationship.swf_asset
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2010-10-09 09:22:40 -07:00
|
|
|
def parent_swf_asset_relationships_to_update=(rels)
|
|
|
|
@parent_swf_asset_relationships_to_update = rels
|
|
|
|
end
|
|
|
|
|
2010-11-06 08:52:58 -07:00
|
|
|
def self.all_by_ids_or_children(ids, swf_assets)
|
|
|
|
swf_asset_ids = []
|
|
|
|
swf_assets_by_id = {}
|
|
|
|
swf_assets_by_parent_id = {}
|
|
|
|
swf_assets.each do |swf_asset|
|
|
|
|
id = swf_asset.id
|
|
|
|
swf_assets_by_id[id] = swf_asset
|
|
|
|
swf_asset_ids << id
|
|
|
|
end
|
|
|
|
SwfAsset.select('id, parent_id').object_assets.joins(:object_asset_relationships).
|
|
|
|
where(SwfAsset.arel_table[:id].in(swf_asset_ids)).each do |row|
|
|
|
|
item_id = row.parent_id.to_i
|
|
|
|
swf_assets_by_parent_id[item_id] ||= []
|
|
|
|
swf_assets_by_parent_id[item_id] << swf_assets_by_id[row.id.to_i]
|
|
|
|
ids << item_id
|
|
|
|
end
|
|
|
|
find(ids).tap do |items|
|
|
|
|
items.each do |item|
|
|
|
|
swf_assets = swf_assets_by_parent_id[item.id]
|
|
|
|
if swf_assets
|
|
|
|
swf_assets.each do |swf_asset|
|
|
|
|
swf_asset.item = item
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2010-10-07 07:46:23 -07:00
|
|
|
def self.collection_from_pet_type_and_registries(pet_type, info_registry, asset_registry)
|
|
|
|
# bear in mind that registries are arrays with many nil elements,
|
|
|
|
# due to how the parser works
|
|
|
|
items = {}
|
|
|
|
item_ids = []
|
|
|
|
info_registry.each do |info|
|
2010-10-09 08:06:59 -07:00
|
|
|
if info && info[:is_compatible]
|
2010-10-07 07:46:23 -07:00
|
|
|
item_ids << info[:obj_info_id].to_i
|
|
|
|
end
|
|
|
|
end
|
|
|
|
existing_relationships_by_item_id_and_swf_asset_id = {}
|
|
|
|
existing_items = Item.find_all_by_id(item_ids, :include => :parent_swf_asset_relationships)
|
|
|
|
existing_items.each do |item|
|
|
|
|
items[item.id] = item
|
|
|
|
relationships_by_swf_asset_id = {}
|
|
|
|
item.parent_swf_asset_relationships.each do |relationship|
|
|
|
|
relationships_by_swf_asset_id[relationship.swf_asset_id] = relationship
|
|
|
|
end
|
|
|
|
existing_relationships_by_item_id_and_swf_asset_id[item.id] =
|
|
|
|
relationships_by_swf_asset_id
|
|
|
|
end
|
|
|
|
swf_asset_ids = []
|
|
|
|
asset_registry.each_with_index do |asset_data, index|
|
|
|
|
swf_asset_ids << index if asset_data
|
|
|
|
end
|
2010-11-20 07:35:59 -08:00
|
|
|
existing_swf_assets = SwfAsset.object_assets.find_all_by_id swf_asset_ids
|
2010-10-07 07:46:23 -07:00
|
|
|
existing_swf_assets_by_id = {}
|
|
|
|
existing_swf_assets.each do |swf_asset|
|
|
|
|
existing_swf_assets_by_id[swf_asset.id] = swf_asset
|
|
|
|
end
|
|
|
|
relationships_by_item_id = {}
|
|
|
|
asset_registry.each do |asset_data|
|
|
|
|
if asset_data
|
|
|
|
item_id = asset_data[:obj_info_id].to_i
|
2010-10-09 08:06:59 -07:00
|
|
|
next unless item_ids.include?(item_id) # skip incompatible
|
2010-10-07 07:46:23 -07:00
|
|
|
item = items[item_id]
|
|
|
|
unless item
|
|
|
|
item = Item.new
|
|
|
|
item.id = item_id
|
|
|
|
items[item_id] = item
|
|
|
|
end
|
|
|
|
item.origin_registry_info = info_registry[item.id]
|
2010-10-09 09:22:40 -07:00
|
|
|
item.current_body_id = pet_type.body_id
|
2010-10-07 07:46:23 -07:00
|
|
|
swf_asset_id = asset_data[:asset_id].to_i
|
2010-10-09 09:22:40 -07:00
|
|
|
swf_asset = existing_swf_assets_by_id[swf_asset_id]
|
2010-10-07 07:46:23 -07:00
|
|
|
unless swf_asset
|
|
|
|
swf_asset = SwfAsset.new
|
|
|
|
swf_asset.id = swf_asset_id
|
|
|
|
end
|
|
|
|
swf_asset.origin_object_data = asset_data
|
|
|
|
swf_asset.origin_pet_type = pet_type
|
|
|
|
relationship = existing_relationships_by_item_id_and_swf_asset_id[item.id][swf_asset_id] rescue nil
|
|
|
|
unless relationship
|
|
|
|
relationship = ParentSwfAssetRelationship.new
|
|
|
|
relationship.parent_id = item.id
|
|
|
|
relationship.swf_asset_type = SwfAssetType
|
|
|
|
relationship.swf_asset_id = swf_asset.id
|
|
|
|
end
|
2010-10-10 19:18:42 -07:00
|
|
|
relationship.object_asset = swf_asset
|
2010-10-07 07:46:23 -07:00
|
|
|
relationships_by_item_id[item_id] ||= []
|
|
|
|
relationships_by_item_id[item_id] << relationship
|
|
|
|
end
|
|
|
|
end
|
|
|
|
relationships_by_item_id.each do |item_id, relationships|
|
2010-10-09 09:22:40 -07:00
|
|
|
items[item_id].parent_swf_asset_relationships_to_update = relationships
|
2010-10-07 07:46:23 -07:00
|
|
|
end
|
|
|
|
items.values
|
|
|
|
end
|
|
|
|
|
2010-11-25 08:10:21 -08:00
|
|
|
class << self
|
|
|
|
MALL_HOST = 'ncmall.neopets.com'
|
|
|
|
MALL_MAIN_PATH = '/mall/shop.phtml'
|
2010-11-27 15:41:06 -08:00
|
|
|
MALL_CATEGORY_PATH = '/mall/ajax/load_page.phtml'
|
|
|
|
MALL_CATEGORY_QUERY = 'type=browse&cat={cat}&lang=en'
|
2010-11-25 08:10:21 -08:00
|
|
|
MALL_CATEGORY_TRIGGER = /load_items_pane\("browse", ([0-9]+)\);/
|
|
|
|
MALL_JSON_ITEM_DATA_KEY = 'object_data'
|
|
|
|
MALL_ITEM_URL_TEMPLATE = 'http://images.neopets.com/items/%s.gif'
|
|
|
|
|
|
|
|
MALL_MAIN_URI = Addressable::URI.new :scheme => 'http',
|
|
|
|
:host => MALL_HOST, :path => MALL_MAIN_PATH
|
|
|
|
MALL_CATEGORY_URI = Addressable::URI.new :scheme => 'http',
|
2010-11-27 15:41:06 -08:00
|
|
|
:host => MALL_HOST, :path => MALL_CATEGORY_PATH,
|
|
|
|
:query => MALL_CATEGORY_QUERY
|
2010-11-25 08:10:21 -08:00
|
|
|
MALL_CATEGORY_TEMPLATE = Addressable::Template.new MALL_CATEGORY_URI
|
|
|
|
|
|
|
|
def spider_mall!
|
|
|
|
# Load the mall HTML, scan it for category onclicks
|
|
|
|
items = {}
|
|
|
|
spider_request(MALL_MAIN_URI).scan(MALL_CATEGORY_TRIGGER) do |match|
|
|
|
|
# Plug the category ID into the URI for that category's JSON document
|
|
|
|
uri = MALL_CATEGORY_TEMPLATE.expand :cat => match[0]
|
|
|
|
begin
|
|
|
|
# Load up that JSON and send it off to be parsed
|
|
|
|
puts "Loading #{uri}..."
|
|
|
|
category_items = spider_mall_category(spider_request(uri))
|
|
|
|
puts "...found #{category_items.size} items"
|
|
|
|
items.merge!(category_items)
|
|
|
|
rescue SpiderJSONError => e
|
|
|
|
# If there was a parsing error, add where it came from
|
|
|
|
Rails.logger.warn "Error parsing JSON at #{uri}, skipping: #{e.message}"
|
|
|
|
end
|
|
|
|
end
|
2010-12-05 13:06:48 -08:00
|
|
|
puts "#{items.size} items found"
|
|
|
|
all_item_ids = items.keys
|
|
|
|
# Find which of these already exist but aren't marked as sold_in_mall so
|
|
|
|
# we can update them as being sold
|
|
|
|
Item.not_sold_in_mall.where(:id => items.keys).select([:id, :name]).each do |item|
|
|
|
|
items.delete(item.id)
|
|
|
|
item.sold_in_mall = true
|
|
|
|
item.save
|
|
|
|
puts "#{item.name} (#{item.id}) now in mall, updated"
|
|
|
|
end
|
|
|
|
# Find items marked as sold_in_mall so we can skip those we just found
|
|
|
|
# if they already are properly marked, and mark those that we didn't just
|
|
|
|
# find as no longer sold_in_mall
|
|
|
|
Item.sold_in_mall.select([:id, :name]).each do |item|
|
|
|
|
if all_item_ids.include?(item.id)
|
|
|
|
items.delete(item.id)
|
|
|
|
else
|
|
|
|
item.sold_in_mall = false
|
|
|
|
item.save
|
|
|
|
puts "#{item.name} (#{item.id}) no longer in mall, removed sold_in_mall status"
|
|
|
|
end
|
|
|
|
end
|
2010-11-25 08:10:21 -08:00
|
|
|
puts "#{items.size} new items"
|
|
|
|
items.each do |item_id, item|
|
|
|
|
item.save
|
|
|
|
puts "Saved #{item.name} (#{item_id})"
|
|
|
|
end
|
|
|
|
items
|
|
|
|
end
|
|
|
|
|
2010-11-27 15:41:06 -08:00
|
|
|
def spider_mall_assets!(limit)
|
|
|
|
items = self.select([arel_table[:id], arel_table[:name]]).sold_in_mall.spidered_longest_ago.limit(limit).all
|
|
|
|
puts "- #{items.size} items need asset spidering"
|
|
|
|
AssetStrategy.build_strategies
|
|
|
|
items.each do |item|
|
|
|
|
AssetStrategy.spider item
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def spider_request(uri)
|
|
|
|
begin
|
|
|
|
response = Net::HTTP.get_response uri
|
|
|
|
rescue SocketError => e
|
|
|
|
raise SpiderHTTPError, "Error loading #{uri}: #{e.message}"
|
|
|
|
end
|
|
|
|
unless response.is_a? Net::HTTPOK
|
|
|
|
raise SpiderHTTPError, "Error loading #{uri}: Response was a #{response.class}"
|
|
|
|
end
|
|
|
|
response.body
|
|
|
|
end
|
|
|
|
|
2010-11-25 08:10:21 -08:00
|
|
|
private
|
|
|
|
|
2010-11-27 15:41:06 -08:00
|
|
|
class AssetStrategy
|
|
|
|
Strategies = {}
|
|
|
|
|
|
|
|
MALL_ASSET_PATH = '/mall/ajax/get_item_assets.phtml'
|
|
|
|
MALL_ASSET_QUERY = 'pet={pet_name}&oii={item_id}'
|
|
|
|
MALL_ASSET_URI = Addressable::URI.new :scheme => 'http',
|
|
|
|
:host => MALL_HOST, :path => MALL_ASSET_PATH,
|
|
|
|
:query => MALL_ASSET_QUERY
|
|
|
|
MALL_ASSET_TEMPLATE = Addressable::Template.new MALL_ASSET_URI
|
|
|
|
|
|
|
|
def initialize(name, options)
|
|
|
|
@name = name
|
|
|
|
@pass = options[:pass]
|
|
|
|
@complete = options[:complete]
|
|
|
|
@pet_types = options[:pet_types]
|
|
|
|
end
|
|
|
|
|
|
|
|
def spider(item)
|
|
|
|
puts " - Using #{@name} strategy"
|
|
|
|
exit = false
|
|
|
|
@pet_types.each do |pet_type|
|
|
|
|
swf_assets = load_for_pet_type(item, pet_type)
|
|
|
|
if swf_assets
|
|
|
|
contains_body_specific_assets = false
|
|
|
|
swf_assets.each do |swf_asset|
|
|
|
|
if swf_asset.body_specific?
|
|
|
|
contains_body_specific_assets = true
|
|
|
|
break
|
|
|
|
end
|
|
|
|
end
|
|
|
|
if contains_body_specific_assets
|
|
|
|
if @pass
|
|
|
|
Strategies[@pass].spider(item) unless @pass == :exit
|
|
|
|
exit = true
|
|
|
|
break
|
|
|
|
end
|
|
|
|
else
|
|
|
|
# if all are universal, no need to spider more
|
|
|
|
puts " - No body specific assets; moving on"
|
|
|
|
exit = true
|
|
|
|
break
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
if !exit && @complete && @complete != :exit
|
|
|
|
Strategies[@complete].spider(item)
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
private
|
|
|
|
|
|
|
|
def load_for_pet_type(item, pet_type, banned_pet_ids=[])
|
|
|
|
pet_id = pet_type.pet_id
|
|
|
|
pet_name = pet_type.pet_name
|
|
|
|
pet = Pet.load(pet_name)
|
|
|
|
if pet.pet_type == pet_type
|
|
|
|
swf_assets = load_for_pet_name(item, pet_type, pet_name)
|
|
|
|
if swf_assets
|
|
|
|
puts " - Modeled with #{pet_name}, saved assets (#{swf_assets.map(&:id).join(', ')})"
|
|
|
|
else
|
|
|
|
puts " - Item #{item.name} does not fit #{pet_name}"
|
|
|
|
end
|
|
|
|
return swf_assets
|
|
|
|
else
|
|
|
|
puts " - Pet #{pet_name} is pet type \##{pet.pet_type_id}, not \##{pet_type.id}; saving it and loading new pet"
|
|
|
|
pet.save
|
|
|
|
banned_pet_ids << pet_id
|
|
|
|
new_pet = pet_type.pets.select([:id, :name]).where(Pet.arel_table[:id].not_in(banned_pet_ids)).first
|
|
|
|
if new_pet
|
|
|
|
pet_type.pet_id = new_pet.id
|
|
|
|
pet_type.pet_name = new_pet.name
|
|
|
|
load_for_pet_type(item, pet_type, banned_pet_ids)
|
|
|
|
else
|
|
|
|
puts " - We have no more pets of type \##{pet_type.id}. Skipping"
|
|
|
|
return nil
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def load_for_pet_name(item, pet_type, pet_name)
|
|
|
|
uri = MALL_ASSET_TEMPLATE.
|
|
|
|
expand(
|
|
|
|
:item_id => item.id,
|
|
|
|
:pet_name => pet_name
|
|
|
|
)
|
|
|
|
raw_data = Item.spider_request(uri)
|
|
|
|
data = JSON.parse(raw_data)
|
|
|
|
item_id_key = item.id.to_s
|
|
|
|
if !data.empty? && data[item_id_key] && data[item_id_key]['asset_data']
|
|
|
|
data[item_id_key]['asset_data'].map do |asset_id_str, asset_data|
|
|
|
|
item.zones_restrict = asset_data['restrict']
|
|
|
|
item.save
|
|
|
|
swf_asset = SwfAsset.find_or_initialize_by_type_and_id(SwfAssetType, asset_id_str.to_i)
|
|
|
|
swf_asset.type = SwfAssetType
|
|
|
|
swf_asset.body_id = pet_type.body_id
|
|
|
|
swf_asset.mall_data = asset_data
|
|
|
|
item.swf_assets << swf_asset unless item.swf_assets.include? swf_asset
|
|
|
|
swf_asset.save
|
|
|
|
swf_asset
|
|
|
|
end
|
|
|
|
else
|
|
|
|
nil
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
class << self
|
|
|
|
def add_strategy(name, options)
|
|
|
|
Strategies[name] = new(name, options)
|
|
|
|
end
|
|
|
|
|
|
|
|
def add_cascading_strategy(name, options)
|
|
|
|
pet_type_groups = options[:pet_types]
|
|
|
|
pet_type_group_names = pet_type_groups.keys
|
|
|
|
pet_type_group_names.each_with_index do |pet_type_group_name, i|
|
|
|
|
remaining_pet_types = pet_type_groups[pet_type_group_name]
|
|
|
|
first_pet_type = [remaining_pet_types.slice!(0)]
|
|
|
|
cascade_name = "#{name}_cascade"
|
|
|
|
next_name = pet_type_group_names[i + 1]
|
|
|
|
next_name = next_name ? "group_#{next_name}" : options[:complete]
|
|
|
|
first_strategy_options = {:complete => next_name, :pass => :exit,
|
|
|
|
:pet_types => first_pet_type}
|
|
|
|
unless remaining_pet_types.empty?
|
|
|
|
first_strategy_options[:pass] = cascade_name
|
|
|
|
add_strategy cascade_name, :complete => :exit,
|
|
|
|
:pet_types => remaining_pet_types
|
|
|
|
end
|
|
|
|
add_strategy name, first_strategy_options
|
|
|
|
name = next_name
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def spider(item)
|
|
|
|
puts "- Spidering for #{item.name}"
|
|
|
|
Strategies[:start].spider(item)
|
|
|
|
item.last_spidered = Time.now
|
|
|
|
item.save
|
|
|
|
puts "- #{item.name} done spidering, saved last spidered timestamp"
|
|
|
|
end
|
|
|
|
|
|
|
|
def build_strategies
|
|
|
|
if Strategies.empty?
|
|
|
|
pet_type_t = PetType.arel_table
|
|
|
|
require 'pet' # FIXME: console is whining when i don't do this
|
|
|
|
pet_t = Pet.arel_table
|
|
|
|
pet_types = PetType.select([pet_type_t[:id], pet_type_t[:body_id], "#{Pet.table_name}.id as pet_id, #{Pet.table_name}.name as pet_name"]).
|
|
|
|
joins(:pets).group(pet_type_t[:id])
|
|
|
|
remaining_standard_pet_types = pet_types.single_standard_color.order(:species_id)
|
|
|
|
first_standard_pet_type = [remaining_standard_pet_types.slice!(0)]
|
|
|
|
|
|
|
|
add_strategy :start, :pass => :remaining_standard, :complete => :first_nonstandard_color,
|
|
|
|
:pet_types => first_standard_pet_type
|
|
|
|
|
|
|
|
add_strategy :remaining_standard, :complete => :exit,
|
|
|
|
:pet_types => remaining_standard_pet_types
|
|
|
|
|
|
|
|
add_cascading_strategy :first_nonstandard_color, :complete => :remaining_standard,
|
|
|
|
:pet_types => pet_types.select(pet_type_t[:color_id]).nonstandard_colors.all.group_by(&:color_id)
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2010-11-25 08:10:21 -08:00
|
|
|
def spider_mall_category(json)
|
|
|
|
begin
|
|
|
|
items_data = JSON.parse(json)[MALL_JSON_ITEM_DATA_KEY]
|
|
|
|
unless items_data
|
|
|
|
raise SpiderJSONError, "Missing key #{MALL_JSON_ITEM_DATA_KEY}"
|
|
|
|
end
|
|
|
|
rescue Exception => e
|
|
|
|
# Catch both errors parsing JSON and the missing key
|
|
|
|
raise SpiderJSONError, e.message
|
|
|
|
end
|
|
|
|
items = {}
|
|
|
|
items_data.each do |item_id, item_data|
|
|
|
|
if item_data['isWearable'] == 1
|
|
|
|
relevant_item_data = item_data.slice('name', 'description', 'price')
|
|
|
|
item = Item.new relevant_item_data
|
|
|
|
item.id = item_data['id']
|
|
|
|
item.thumbnail_url = sprintf(MALL_ITEM_URL_TEMPLATE, item_data['imageFile'])
|
|
|
|
item.sold_in_mall = true
|
|
|
|
items[item.id] = item
|
|
|
|
end
|
|
|
|
end
|
|
|
|
items
|
|
|
|
end
|
|
|
|
|
|
|
|
class SpiderError < RuntimeError;end
|
|
|
|
class SpiderHTTPError < SpiderError;end
|
|
|
|
class SpiderJSONError < SpiderError;end
|
|
|
|
end
|
|
|
|
|
2010-05-15 08:38:45 -07:00
|
|
|
private
|
|
|
|
|
2010-06-22 09:42:25 -07:00
|
|
|
SearchFilterScopes = []
|
2010-11-18 20:21:03 -08:00
|
|
|
LimitedSearchFilters = []
|
|
|
|
|
|
|
|
def self.search_filter(name, options={}, &block)
|
|
|
|
assume_complement = options.delete(:assume_complement) || true
|
|
|
|
name = name.to_s
|
|
|
|
SearchFilterScopes << name
|
|
|
|
LimitedSearchFilters << name if options[:limit]
|
|
|
|
(class << self; self; end).instance_eval do
|
|
|
|
if options[:full]
|
|
|
|
define_method "search_filter_#{name}", &options[:full]
|
|
|
|
else
|
|
|
|
if assume_complement
|
|
|
|
define_method "search_filter_not_#{name}", &Item.search_filter_block(options, false, &block)
|
|
|
|
end
|
|
|
|
define_method "search_filter_#{name}", &Item.search_filter_block(options, true, &block)
|
|
|
|
end
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def self.single_search_filter(name, options={}, &block)
|
|
|
|
options[:assume_complement] = false
|
|
|
|
search_filter name, options, &block
|
|
|
|
end
|
2010-06-22 09:42:25 -07:00
|
|
|
|
2010-11-18 20:21:03 -08:00
|
|
|
def self.search_filter_block(options, positive)
|
|
|
|
Proc.new { |str, scope|
|
2010-06-22 09:42:25 -07:00
|
|
|
condition = yield(str)
|
2010-11-18 20:21:03 -08:00
|
|
|
condition = "!(#{condition.to_sql})" unless positive
|
|
|
|
scope = scope.send(options[:scope]) if options[:scope]
|
|
|
|
scope.where(condition)
|
2010-06-22 09:42:25 -07:00
|
|
|
}
|
|
|
|
end
|
|
|
|
|
|
|
|
search_filter :name do |name|
|
|
|
|
arel_table[:name].matches("%#{name}%")
|
|
|
|
end
|
|
|
|
|
|
|
|
search_filter :description do |description|
|
|
|
|
arel_table[:description].matches("%#{description}%")
|
|
|
|
end
|
|
|
|
|
2010-10-03 17:14:48 -07:00
|
|
|
ADJECTIVE_FILTERS = {
|
|
|
|
'nc' => arel_table[:rarity_index].in(NCRarities),
|
|
|
|
'pb' => arel_table[:description].eq(PaintbrushSetDescription)
|
|
|
|
}
|
|
|
|
search_filter :is do |adjective|
|
|
|
|
filter = ADJECTIVE_FILTERS[adjective]
|
|
|
|
unless filter
|
2010-11-16 20:32:53 -08:00
|
|
|
raise SearchError,
|
2010-10-03 17:14:48 -07:00
|
|
|
"We don't know how an item can be \"#{adjective}\". " +
|
|
|
|
"Did you mean is:nc or is:pb?"
|
|
|
|
end
|
|
|
|
filter
|
2010-07-10 09:42:18 -07:00
|
|
|
end
|
|
|
|
|
2010-06-22 09:42:25 -07:00
|
|
|
search_filter :only do |species_name|
|
2010-11-18 20:25:34 -08:00
|
|
|
begin
|
|
|
|
id = Species.require_by_name(species_name).id
|
|
|
|
rescue Species::NotFound => e
|
|
|
|
raise SearchError, e.message
|
|
|
|
end
|
2010-06-22 09:42:25 -07:00
|
|
|
arel_table[:species_support_ids].eq(id.to_s)
|
|
|
|
end
|
|
|
|
|
|
|
|
search_filter :species do |species_name|
|
2010-11-18 20:25:34 -08:00
|
|
|
begin
|
|
|
|
id = Species.require_by_name(species_name).id
|
|
|
|
rescue Species::NotFound => e
|
|
|
|
raise SearchError, e.message
|
|
|
|
end
|
2010-06-22 09:42:25 -07:00
|
|
|
ids = arel_table[:species_support_ids]
|
2010-11-16 20:32:53 -08:00
|
|
|
ids.eq('').or(ids.matches_any([
|
2010-06-22 09:42:25 -07:00
|
|
|
id,
|
|
|
|
"#{id},%",
|
|
|
|
"%,#{id},%",
|
|
|
|
"%,#{id}"
|
2010-11-16 20:32:53 -08:00
|
|
|
]))
|
2010-06-22 09:42:25 -07:00
|
|
|
end
|
|
|
|
|
2010-11-18 20:21:03 -08:00
|
|
|
single_search_filter :type, {:limit => true, :scope => :join_swf_assets} do |zone_set_name|
|
2010-06-22 09:42:25 -07:00
|
|
|
zone_set = Zone::ItemZoneSets[zone_set_name]
|
2010-11-16 20:32:53 -08:00
|
|
|
raise SearchError, "Type \"#{zone_set_name}\" does not exist" unless zone_set
|
2010-06-22 09:42:25 -07:00
|
|
|
SwfAsset.arel_table[:zone_id].in(zone_set.map(&:id))
|
|
|
|
end
|
|
|
|
|
2010-11-18 20:21:03 -08:00
|
|
|
single_search_filter :not_type, :full => lambda { |zone_set_name, scope|
|
|
|
|
zone_set = Zone::ItemZoneSets[zone_set_name]
|
|
|
|
raise SearchError, "Type \"#{zone_set_name}\" does not exist" unless zone_set
|
|
|
|
psa = ParentSwfAssetRelationship.arel_table.alias
|
|
|
|
sa = SwfAsset.arel_table.alias
|
|
|
|
# Join to SWF assets, including the zone condition in the join so that
|
|
|
|
# SWFs that don't match end up being NULL rows. Then we take the max SWF
|
|
|
|
# asset ID, which is NULL if and only if there are no rows that matched
|
|
|
|
# the zone requirement. If that max was NULL, return the object.
|
2010-11-19 13:31:58 -08:00
|
|
|
item_ids = select(arel_table[:id]).joins(
|
2010-11-18 20:21:03 -08:00
|
|
|
"LEFT JOIN #{ParentSwfAssetRelationship.table_name} #{psa.name} ON " +
|
|
|
|
psa[:swf_asset_type].eq(SwfAssetType)
|
|
|
|
.and(psa[:parent_id].eq(arel_table[:id]))
|
|
|
|
.to_sql
|
|
|
|
).
|
|
|
|
joins(
|
|
|
|
"LEFT JOIN #{SwfAsset.table_name} #{sa.name} ON " +
|
|
|
|
sa[:type].eq(SwfAssetType)
|
|
|
|
.and(sa[:id].eq(psa[:swf_asset_id]))
|
|
|
|
.and(sa[:zone_id].in(zone_set.map(&:id)))
|
|
|
|
.to_sql
|
|
|
|
).
|
|
|
|
group("#{table_name}.id").
|
2010-11-19 13:31:58 -08:00
|
|
|
having("MAX(#{sa.name}.id) IS NULL"). # SwfAsset.arel_table[:id].maximum has no #eq
|
|
|
|
map(&:id)
|
|
|
|
scope.where(arel_table[:id].in(item_ids))
|
2010-11-18 20:21:03 -08:00
|
|
|
}
|
|
|
|
|
2010-05-15 08:38:45 -07:00
|
|
|
class Condition < String
|
2010-11-18 20:21:03 -08:00
|
|
|
attr_accessor :filter
|
|
|
|
|
|
|
|
def initialize
|
|
|
|
@positive = true
|
|
|
|
end
|
|
|
|
|
|
|
|
def filter?
|
|
|
|
!@filter.nil?
|
|
|
|
end
|
|
|
|
|
2010-06-22 09:42:25 -07:00
|
|
|
def to_filter!
|
|
|
|
@filter = self.clone
|
2010-05-15 08:38:45 -07:00
|
|
|
self.replace ''
|
|
|
|
end
|
|
|
|
|
2010-05-15 09:43:54 -07:00
|
|
|
def negate!
|
2010-11-18 20:21:03 -08:00
|
|
|
@positive = !@positive
|
2010-05-15 09:43:54 -07:00
|
|
|
end
|
|
|
|
|
2010-05-15 08:38:45 -07:00
|
|
|
def narrow(scope)
|
2010-06-22 09:42:25 -07:00
|
|
|
if SearchFilterScopes.include?(filter)
|
2010-11-18 20:21:03 -08:00
|
|
|
polarized_filter = @positive ? filter : "not_#{filter}"
|
|
|
|
Item.send("search_filter_#{polarized_filter}", self, scope)
|
2010-05-15 12:00:53 -07:00
|
|
|
else
|
2010-11-16 20:32:53 -08:00
|
|
|
raise SearchError, "Filter #{filter} does not exist"
|
2010-05-15 08:38:45 -07:00
|
|
|
end
|
2010-06-22 09:42:25 -07:00
|
|
|
end
|
|
|
|
|
|
|
|
def filter
|
|
|
|
@filter || 'name'
|
2010-05-15 08:38:45 -07:00
|
|
|
end
|
|
|
|
|
|
|
|
def inspect
|
2010-06-22 09:42:25 -07:00
|
|
|
@filter ? "#{@filter}:#{super}" : super
|
2010-05-15 08:38:45 -07:00
|
|
|
end
|
|
|
|
end
|
2010-11-16 20:32:53 -08:00
|
|
|
|
|
|
|
class SearchError < ArgumentError;end
|
2010-05-14 15:41:40 -07:00
|
|
|
end
|