标签:attr sso 证明 reverse padding return 主机名 failed elf
全文检索不同于特定字段的模糊查询,使用全文检索的效率再高,并且能够对于中文进行分词处理。
haystack:全文检索框架,支持whoosh、solr、Xaplan、Elasticsearc四种全文检索引擎
whoosh:纯python编写的全文搜索引擎,虽然 性能比不上sphinx、xapian、elasticsearc等,但是无二进制包,程序不会莫名其妙的崩溃,对于小型的站点,whoosh已经足够使用,
安装包:
pip install django-haystack
pip install whoosh
pip install jieba
修改settings文件
INSTALLWS_APPS = [
‘haystack‘, # 全文检索框架
]
?
?
全文检索框架的配置
HAYSTACK_CONNECTIONS = {
‘default‘: {
# 使用whoosh搜索引擎
‘ENGINE‘: ‘goods.whoosh_cn_backend.WhooshEngine‘,
# 索引文件的路径
‘PATH‘: os.path.join(BASE_DIR, ‘whoosh_index‘),
},
}
?
# 当添加、修改、删除数据时,自动生成索引
HAYSTACK_SIGNAL_PROCESSOR = ‘haystack.signals.RealtimeSignalProcessor‘
?
# 设置每页显示的数目,默认为20,可以自己修改
HAYSTACK_SEARCH_RESULTS_PER_PAGE = 8
在要检索的app下面,创建search_indexes.py文件
# 定义索引类
from haystack import indexes
from .models import Goods
?
?
# 索引类名格式:模型类名+Index
class GoodsIndex(indexes.SearchIndex, indexes.Indexable):
"""
?
"""
# 索引字段:use_template 指定根据表中的哪些字段 建立索引文件
# 把说明放在一个文件中
text = indexes.CharField(document=True, use_template=True)
# 建立检索字段,model_attr模型属性,如果需要多字段的话,在这里添加需要检索的字段
goods_name = indexes.NgramField(model_attr="goods_name")
?
def get_model(self):
return Goods # 返回的模型类
?
def index_queryset(self, using=None):
return self.get_model().objects.all()
?
在templates下面创建如下文件夹 search/indexes/goods,在这下面创建goods_text.txt(goods是需要检索的模型类的小写)
注:名称是固定的,不可随意更改
在goods_test.txt里面写入需要检索的字段
# 指定根据表中的哪些字段建立索引数据
{{object.goods_name}} # 根据商品的名称建立索引
?
进入到项目所在的目录,创建索引文件:python manage.py rebuild_index
html下搜索框固定设置
<form action="./search" method="get"><--form的method必须为get-->
{% csrf_token %}
<input type="text" placeholder="搜索品牌 店铺..." name="q"> <--inpu的name必须为q-->
<input type="submit" value="搜索" >
</form>
配置项目下的urls
from django.urls import path, include
?
urlpatterns = [
path(‘search/‘, include(‘haystack.urls‘)) # 全文检过框架
]
搜索出来的结果,haystack会把搜索结果传递给templates/search目录下的search.html,所以需要在templates的search文件夹下创建search.html文件。传递的上下文包括:
query: 搜索的关键字
page:当前页的page对象
serchResult类的实例对象,对象的属性是object
paginator:分页paginator对象
# 设置每页显示的数目,默认为20,可以自己修改
HAYSTACK_SEARCH_RESULTS_PER_PAGE = 8
配置中文分词器,这里用到的模块为jieba,文件名为:tokenizer.py,把本文件放在与search_indexes.py同目录下,我这里放在了goods文件夹下
from jieba import cut_for_search
from whoosh.analysis import Tokenizer, Token
?
?
class ChineseTokenizer(Tokenizer):
def __call__(self, value, positions=False, chars=False,
keeporiginal=False, removestops=True,
start_pos=0, start_char=0, mode=‘‘, **kwargs):
?
t = Token(positions, chars, removestops=removestops, mode=mode,
**kwargs)
# seglist = cut(value, cut_all=False) # (精确模式)使用结巴分词库进行分词
seglist = cut_for_search(value) # (搜索引擎模式) 使用结巴分词库进行分词
for w in seglist:
t.original = t.text = w
t.boost = 1.0
if positions:
t.pos = start_pos + value.find(w)
if chars:
t.startchar = start_char + value.find(w)
t.endchar = start_char + value.find(w) + len(w)
yield t # 通过生成器返回每个分词的结果token
?
?
def ChineseAnalyzer():
return ChineseTokenizer()
?
中文搜索引擎配置,文件名:whoosh_cn_backend.py,把本文件放在与search_indexes.py同目录下,我这里放在了goods文件夹下,修改172行的app名
# encoding: utf-8
?
from __future__ import absolute_import, division, print_function, unicode_literals
?
import json
import os
import re
import shutil
import threading
import warnings
?
from django.conf import settings
from django.core.exceptions import ImproperlyConfigured
from django.utils import six
from django.utils.datetime_safe import datetime
from django.utils.encoding import force_text
?
from haystack.backends import BaseEngine, BaseSearchBackend, BaseSearchQuery, EmptyResults, log_query
from haystack.constants import DJANGO_CT, DJANGO_ID, ID
from haystack.exceptions import MissingDependency, SearchBackendError, SkipDocument
from haystack.inputs import Clean, Exact, PythonData, Raw
from haystack.models import SearchResult
from haystack.utils import log as logging
from haystack.utils import get_identifier, get_model_ct
from haystack.utils.app_loading import haystack_get_model
?
try:
import whoosh
except ImportError:
raise MissingDependency(
"The ‘whoosh‘ backend requires the installation of ‘Whoosh‘. Please refer to the documentation.")
?
# Handle minimum requirement.
if not hasattr(whoosh, ‘__version__‘) or whoosh.__version__ < (2, 5, 0):
raise MissingDependency("The ‘whoosh‘ backend requires version 2.5.0 or greater.")
?
# Bubble up the correct error.
from whoosh import index
from whoosh.analysis import StemmingAnalyzer
from whoosh.fields import ID as WHOOSH_ID
from whoosh.fields import BOOLEAN, DATETIME, IDLIST, KEYWORD, NGRAM, NGRAMWORDS, NUMERIC, Schema, TEXT
from whoosh.filedb.filestore import FileStorage, RamStorage
from whoosh.highlight import highlight as whoosh_highlight
from whoosh.highlight import ContextFragmenter, HtmlFormatter
from whoosh.qparser import QueryParser
from whoosh.searching import ResultsPage
from whoosh.writing import AsyncWriter
?
DATETIME_REGEX = re.compile(
‘^(?P<year>\d{4})-(?P<month>\d{2})-(?P<day>\d{2})T(?P<hour>\d{2}):(?P<minute>\d{2}):(?P<second>\d{2})(\.\d{3,6}Z?)?$‘)
LOCALS = threading.local()
LOCALS.RAM_STORE = None
?
?
class WhooshHtmlFormatter(HtmlFormatter):
"""
This is a HtmlFormatter simpler than the whoosh.HtmlFormatter.
We use it to have consistent results across backends. Specifically,
Solr, Xapian and Elasticsearch are using this formatting.
"""
template = ‘<%(tag)s>%(t)s</%(tag)s>‘
?
?
class WhooshSearchBackend(BaseSearchBackend):
# Word reserved by Whoosh for special use.
RESERVED_WORDS = (
‘AND‘,
‘NOT‘,
‘OR‘,
‘TO‘,
)
?
# Characters reserved by Whoosh for special use.
# The ‘\\‘ must come first, so as not to overwrite the other slash replacements.
RESERVED_CHARACTERS = (
‘\\‘, ‘+‘, ‘-‘, ‘&&‘, ‘||‘, ‘!‘, ‘(‘, ‘)‘, ‘{‘, ‘}‘,
‘[‘, ‘]‘, ‘^‘, ‘"‘, ‘~‘, ‘*‘, ‘?‘, ‘:‘, ‘.‘,
)
?
def __init__(self, connection_alias, **connection_options):
super(WhooshSearchBackend, self).__init__(connection_alias, **connection_options)
self.setup_complete = False
self.use_file_storage = True
self.post_limit = getattr(connection_options, ‘POST_LIMIT‘, 128 * 1024 * 1024)
self.path = connection_options.get(‘PATH‘)
?
if connection_options.get(‘STORAGE‘, ‘file‘) != ‘file‘:
self.use_file_storage = False
?
if self.use_file_storage and not self.path:
raise ImproperlyConfigured(
"You must specify a ‘PATH‘ in your settings for connection ‘%s‘." % connection_alias)
?
self.log = logging.getLogger(‘haystack‘)
?
def setup(self):
"""
Defers loading until needed.
"""
from haystack import connections
new_index = False
?
# Make sure the index is there.
if self.use_file_storage and not os.path.exists(self.path):
os.makedirs(self.path)
new_index = True
?
if self.use_file_storage and not os.access(self.path, os.W_OK):
raise IOError("The path to your Whoosh index ‘%s‘ is not writable for the current user/group." % self.path)
?
if self.use_file_storage:
self.storage = FileStorage(self.path)
else:
global LOCALS
?
if getattr(LOCALS, ‘RAM_STORE‘, None) is None:
LOCALS.RAM_STORE = RamStorage()
?
self.storage = LOCALS.RAM_STORE
?
self.content_field_name, self.schema = self.build_schema(
connections[self.connection_alias].get_unified_index().all_searchfields())
self.parser = QueryParser(self.content_field_name, schema=self.schema)
?
if new_index is True:
self.index = self.storage.create_index(self.schema)
else:
try:
self.index = self.storage.open_index(schema=self.schema)
except index.EmptyIndexError:
self.index = self.storage.create_index(self.schema)
?
self.setup_complete = True
?
def build_schema(self, fields):
schema_fields = {
ID: WHOOSH_ID(stored=True, unique=True),
DJANGO_CT: WHOOSH_ID(stored=True),
DJANGO_ID: WHOOSH_ID(stored=True),
}
# Grab the number of keys that are hard-coded into Haystack.
# We‘ll use this to (possibly) fail slightly more gracefully later.
initial_key_count = len(schema_fields)
content_field_name = ‘‘
?
for field_name, field_class in fields.items():
if field_class.is_multivalued:
if field_class.indexed is False:
schema_fields[field_class.index_fieldname] = IDLIST(stored=True, field_boost=field_class.boost)
else:
schema_fields[field_class.index_fieldname] = KEYWORD(stored=True, commas=True, scorable=True,
field_boost=field_class.boost)
elif field_class.field_type in [‘date‘, ‘datetime‘]:
schema_fields[field_class.index_fieldname] = DATETIME(stored=field_class.stored, sortable=True)
elif field_class.field_type == ‘integer‘:
schema_fields[field_class.index_fieldname] = NUMERIC(stored=field_class.stored, numtype=int,
field_boost=field_class.boost)
elif field_class.field_type == ‘float‘:
schema_fields[field_class.index_fieldname] = NUMERIC(stored=field_class.stored, numtype=float,
field_boost=field_class.boost)
elif field_class.field_type == ‘boolean‘:
# Field boost isn‘t supported on BOOLEAN as of 1.8.2.
schema_fields[field_class.index_fieldname] = BOOLEAN(stored=field_class.stored)
elif field_class.field_type == ‘ngram‘:
schema_fields[field_class.index_fieldname] = NGRAM(minsize=3, maxsize=15, stored=field_class.stored,
field_boost=field_class.boost)
elif field_class.field_type == ‘edge_ngram‘:
schema_fields[field_class.index_fieldname] = NGRAMWORDS(minsize=2, maxsize=15, at=‘start‘,
stored=field_class.stored,
field_boost=field_class.boost)
else:
from goods.tokenizer import ChineseAnalyzer
schema_fields[field_class.index_fieldname] = TEXT(stored=True, analyzer=ChineseAnalyzer(),
field_boost=field_class.boost, sortable=True)
if field_class.document is True:
content_field_name = field_class.index_fieldname
schema_fields[field_class.index_fieldname].spelling = True
?
# Fail more gracefully than relying on the backend to die if no fields
# are found.
if len(schema_fields) <= initial_key_count:
raise SearchBackendError(
"No fields were found in any search_indexes. Please correct this before attempting to search.")
?
return (content_field_name, Schema(**schema_fields))
?
def update(self, index, iterable, commit=True):
if not self.setup_complete:
self.setup()
?
self.index = self.index.refresh()
writer = AsyncWriter(self.index)
?
for obj in iterable:
try:
doc = index.full_prepare(obj)
except SkipDocument:
self.log.debug(u"Indexing for object `%s` skipped", obj)
else:
# Really make sure it‘s unicode, because Whoosh won‘t have it any
# other way.
for key in doc:
doc[key] = self._from_python(doc[key])
?
# Document boosts aren‘t supported in Whoosh 2.5.0+.
if ‘boost‘ in doc:
del doc[‘boost‘]
?
try:
writer.update_document(**doc)
except Exception as e:
if not self.silently_fail:
raise
?
# We‘ll log the object identifier but won‘t include the actual object
# to avoid the possibility of that generating encoding errors while
# processing the log message:
self.log.error(u"%s while preparing object for update" % e.__class__.__name__,
exc_info=True, extra={"data": {"index": index,
"object": get_identifier(obj)}})
?
if len(iterable) > 0:
# For now, commit no matter what, as we run into locking issues otherwise.
writer.commit()
?
def remove(self, obj_or_string, commit=True):
if not self.setup_complete:
self.setup()
?
self.index = self.index.refresh()
whoosh_id = get_identifier(obj_or_string)
?
try:
self.index.delete_by_query(q=self.parser.parse(u‘%s:"%s"‘ % (ID, whoosh_id)))
except Exception as e:
if not self.silently_fail:
raise
?
self.log.error("Failed to remove document ‘%s‘ from Whoosh: %s", whoosh_id, e, exc_info=True)
?
def clear(self, models=None, commit=True):
if not self.setup_complete:
self.setup()
?
self.index = self.index.refresh()
?
if models is not None:
assert isinstance(models, (list, tuple))
?
try:
if models is None:
self.delete_index()
else:
models_to_delete = []
?
for model in models:
models_to_delete.append(u"%s:%s" % (DJANGO_CT, get_model_ct(model)))
?
self.index.delete_by_query(q=self.parser.parse(u" OR ".join(models_to_delete)))
except Exception as e:
if not self.silently_fail:
raise
?
if models is not None:
self.log.error("Failed to clear Whoosh index of models ‘%s‘: %s", ‘,‘.join(models_to_delete),
e, exc_info=True)
else:
self.log.error("Failed to clear Whoosh index: %s", e, exc_info=True)
?
def delete_index(self):
# Per the Whoosh mailing list, if wiping out everything from the index,
# it‘s much more efficient to simply delete the index files.
if self.use_file_storage and os.path.exists(self.path):
shutil.rmtree(self.path)
elif not self.use_file_storage:
self.storage.clean()
?
# Recreate everything.
self.setup()
?
def optimize(self):
if not self.setup_complete:
self.setup()
?
self.index = self.index.refresh()
self.index.optimize()
?
def calculate_page(self, start_offset=0, end_offset=None):
# Prevent against Whoosh throwing an error. Requires an end_offset
# greater than 0.
if end_offset is not None and end_offset <= 0:
end_offset = 1
?
# Determine the page.
page_num = 0
?
if end_offset is None:
end_offset = 1000000
?
if start_offset is None:
start_offset = 0
?
page_length = end_offset - start_offset
?
if page_length and page_length > 0:
page_num = int(start_offset / page_length)
?
# Increment because Whoosh uses 1-based page numbers.
page_num += 1
return page_num, page_length
?
def search(self, query_string, sort_by=None, start_offset=0, end_offset=None,
fields=‘‘, highlight=False, facets=None, date_facets=None, query_facets=None,
narrow_queries=None, spelling_query=None, within=None,
dwithin=None, distance_point=None, models=None,
limit_to_registered_models=None, result_class=None, **kwargs):
if not self.setup_complete:
self.setup()
?
# A zero length query should return no results.
if len(query_string) == 0:
return {
‘results‘: [],
‘hits‘: 0,
}
?
query_string = force_text(query_string)
?
# A one-character query (non-wildcard) gets nabbed by a stopwords
# filter and should yield zero results.
if len(query_string) <= 1 and query_string != u‘*‘:
return {
‘results‘: [],
‘hits‘: 0,
}
?
reverse = False
?
if sort_by is not None:
# Determine if we need to reverse the results and if Whoosh can
# handle what it‘s being asked to sort by. Reversing is an
# all-or-nothing action, unfortunately.
sort_by_list = []
reverse_counter = 0
?
for order_by in sort_by:
if order_by.startswith(‘-‘):
reverse_counter += 1
?
if reverse_counter and reverse_counter != len(sort_by):
raise SearchBackendError("Whoosh requires all order_by fields"
" to use the same sort direction")
?
for order_by in sort_by:
if order_by.startswith(‘-‘):
sort_by_list.append(order_by[1:])
?
if len(sort_by_list) == 1:
reverse = True
else:
sort_by_list.append(order_by)
?
if len(sort_by_list) == 1:
reverse = False
?
sort_by = sort_by_list
?
if facets is not None:
warnings.warn("Whoosh does not handle faceting.", Warning, stacklevel=2)
?
if date_facets is not None:
warnings.warn("Whoosh does not handle date faceting.", Warning, stacklevel=2)
?
if query_facets is not None:
warnings.warn("Whoosh does not handle query faceting.", Warning, stacklevel=2)
?
narrowed_results = None
self.index = self.index.refresh()
?
if limit_to_registered_models is None:
limit_to_registered_models = getattr(settings, ‘HAYSTACK_LIMIT_TO_REGISTERED_MODELS‘, True)
?
if models and len(models):
model_choices = sorted(get_model_ct(model) for model in models)
elif limit_to_registered_models:
# Using narrow queries, limit the results to only models handled
# with the current routers.
model_choices = self.build_models_list()
else:
model_choices = []
?
if len(model_choices) > 0:
if narrow_queries is None:
narrow_queries = set()
?
narrow_queries.add(‘ OR ‘.join([‘%s:%s‘ % (DJANGO_CT, rm) for rm in model_choices]))
?
narrow_searcher = None
?
if narrow_queries is not None:
# Potentially expensive? I don‘t see another way to do it in Whoosh...
narrow_searcher = self.index.searcher()
?
for nq in narrow_queries:
recent_narrowed_results = narrow_searcher.search(self.parser.parse(force_text(nq)),
limit=None)
?
if len(recent_narrowed_results) <= 0:
return {
‘results‘: [],
‘hits‘: 0,
}
?
if narrowed_results:
narrowed_results.filter(recent_narrowed_results)
else:
narrowed_results = recent_narrowed_results
?
self.index = self.index.refresh()
?
if self.index.doc_count():
searcher = self.index.searcher()
parsed_query = self.parser.parse(query_string)
?
# In the event of an invalid/stopworded query, recover gracefully.
if parsed_query is None:
return {
‘results‘: [],
‘hits‘: 0,
}
?
page_num, page_length = self.calculate_page(start_offset, end_offset)
?
search_kwargs = {
‘pagelen‘: page_length,
‘sortedby‘: sort_by,
‘reverse‘: reverse,
}
?
# Handle the case where the results have been narrowed.
if narrowed_results is not None:
search_kwargs[‘filter‘] = narrowed_results
?
try:
raw_page = searcher.search_page(
parsed_query,
page_num,
**search_kwargs
)
except ValueError:
if not self.silently_fail:
raise
?
return {
‘results‘: [],
‘hits‘: 0,
‘spelling_suggestion‘: None,
}
?
# Because as of Whoosh 2.5.1, it will return the wrong page of
# results if you request something too high. :(
if raw_page.pagenum < page_num:
return {
‘results‘: [],
‘hits‘: 0,
‘spelling_suggestion‘: None,
}
?
results = self._process_results(raw_page, highlight=highlight, query_string=query_string,
spelling_query=spelling_query, result_class=result_class)
searcher.close()
?
if hasattr(narrow_searcher, ‘close‘):
narrow_searcher.close()
?
return results
else:
if self.include_spelling:
if spelling_query:
spelling_suggestion = self.create_spelling_suggestion(spelling_query)
else:
spelling_suggestion = self.create_spelling_suggestion(query_string)
else:
spelling_suggestion = None
?
return {
‘results‘: [],
‘hits‘: 0,
‘spelling_suggestion‘: spelling_suggestion,
}
?
def more_like_this(self, model_instance, additional_query_string=None,
start_offset=0, end_offset=None, models=None,
limit_to_registered_models=None, result_class=None, **kwargs):
if not self.setup_complete:
self.setup()
?
field_name = self.content_field_name
narrow_queries = set()
narrowed_results = None
self.index = self.index.refresh()
?
if limit_to_registered_models is None:
limit_to_registered_models = getattr(settings, ‘HAYSTACK_LIMIT_TO_REGISTERED_MODELS‘, True)
?
if models and len(models):
model_choices = sorted(get_model_ct(model) for model in models)
elif limit_to_registered_models:
# Using narrow queries, limit the results to only models handled
# with the current routers.
model_choices = self.build_models_list()
else:
model_choices = []
?
if len(model_choices) > 0:
if narrow_queries is None:
narrow_queries = set()
?
narrow_queries.add(‘ OR ‘.join([‘%s:%s‘ % (DJANGO_CT, rm) for rm in model_choices]))
?
if additional_query_string and additional_query_string != ‘*‘:
narrow_queries.add(additional_query_string)
?
narrow_searcher = None
?
if narrow_queries is not None:
# Potentially expensive? I don‘t see another way to do it in Whoosh...
narrow_searcher = self.index.searcher()
?
for nq in narrow_queries:
recent_narrowed_results = narrow_searcher.search(self.parser.parse(force_text(nq)),
limit=None)
?
if len(recent_narrowed_results) <= 0:
return {
‘results‘: [],
‘hits‘: 0,
}
?
if narrowed_results:
narrowed_results.filter(recent_narrowed_results)
else:
narrowed_results = recent_narrowed_results
?
page_num, page_length = self.calculate_page(start_offset, end_offset)
?
self.index = self.index.refresh()
raw_results = EmptyResults()
?
searcher = None
if self.index.doc_count():
query = "%s:%s" % (ID, get_identifier(model_instance))
searcher = self.index.searcher()
parsed_query = self.parser.parse(query)
results = searcher.search(parsed_query)
?
if len(results):
raw_results = results[0].more_like_this(field_name, top=end_offset)
?
# Handle the case where the results have been narrowed.
if narrowed_results is not None and hasattr(raw_results, ‘filter‘):
raw_results.filter(narrowed_results)
?
try:
raw_page = ResultsPage(raw_results, page_num, page_length)
except ValueError:
if not self.silently_fail:
raise
?
return {
‘results‘: [],
‘hits‘: 0,
‘spelling_suggestion‘: None,
}
?
# Because as of Whoosh 2.5.1, it will return the wrong page of
# results if you request something too high. :(
if raw_page.pagenum < page_num:
return {
‘results‘: [],
‘hits‘: 0,
‘spelling_suggestion‘: None,
}
?
results = self._process_results(raw_page, result_class=result_class)
?
if searcher:
searcher.close()
?
if hasattr(narrow_searcher, ‘close‘):
narrow_searcher.close()
?
return results
?
def _process_results(self, raw_page, highlight=False, query_string=‘‘, spelling_query=None, result_class=None):
from haystack import connections
results = []
?
# It‘s important to grab the hits first before slicing. Otherwise, this
# can cause pagination failures.
hits = len(raw_page)
?
if result_class is None:
result_class = SearchResult
?
facets = {}
spelling_suggestion = None
unified_index = connections[self.connection_alias].get_unified_index()
indexed_models = unified_index.get_indexed_models()
?
for doc_offset, raw_result in enumerate(raw_page):
score = raw_page.score(doc_offset) or 0
app_label, model_name = raw_result[DJANGO_CT].split(‘.‘)
additional_fields = {}
model = haystack_get_model(app_label, model_name)
?
if model and model in indexed_models:
for key, value in raw_result.items():
index = unified_index.get_index(model)
string_key = str(key)
?
if string_key in index.fields and hasattr(index.fields[string_key], ‘convert‘):
# Special-cased due to the nature of KEYWORD fields.
if index.fields[string_key].is_multivalued:
if value is None or len(value) is 0:
additional_fields[string_key] = []
else:
additional_fields[string_key] = value.split(‘,‘)
else:
additional_fields[string_key] = index.fields[string_key].convert(value)
else:
additional_fields[string_key] = self._to_python(value)
?
del (additional_fields[DJANGO_CT])
del (additional_fields[DJANGO_ID])
?
if highlight:
sa = StemmingAnalyzer()
formatter = WhooshHtmlFormatter(‘em‘)
terms = [token.text for token in sa(query_string)]
?
whoosh_result = whoosh_highlight(
additional_fields.get(self.content_field_name),
terms,
sa,
ContextFragmenter(),
formatter
)
additional_fields[‘highlighted‘] = {
self.content_field_name: [whoosh_result],
}
?
result = result_class(app_label, model_name, raw_result[DJANGO_ID], score, **additional_fields)
results.append(result)
else:
hits -= 1
?
if self.include_spelling:
if spelling_query:
spelling_suggestion = self.create_spelling_suggestion(spelling_query)
else:
spelling_suggestion = self.create_spelling_suggestion(query_string)
?
return {
‘results‘: results,
‘hits‘: hits,
‘facets‘: facets,
‘spelling_suggestion‘: spelling_suggestion,
}
?
def create_spelling_suggestion(self, query_string):
spelling_suggestion = None
reader = self.index.reader()
corrector = reader.corrector(self.content_field_name)
cleaned_query = force_text(query_string)
?
if not query_string:
return spelling_suggestion
?
# Clean the string.
for rev_word in self.RESERVED_WORDS:
cleaned_query = cleaned_query.replace(rev_word, ‘‘)
?
for rev_char in self.RESERVED_CHARACTERS:
cleaned_query = cleaned_query.replace(rev_char, ‘‘)
?
# Break it down.
query_words = cleaned_query.split()
suggested_words = []
?
for word in query_words:
suggestions = corrector.suggest(word, limit=1)
?
if len(suggestions) > 0:
suggested_words.append(suggestions[0])
?
spelling_suggestion = ‘ ‘.join(suggested_words)
return spelling_suggestion
?
def _from_python(self, value):
"""
Converts Python values to a string for Whoosh.
?
Code courtesy of pysolr.
"""
if hasattr(value, ‘strftime‘):
if not hasattr(value, ‘hour‘):
value = datetime(value.year, value.month, value.day, 0, 0, 0)
elif isinstance(value, bool):
if value:
value = ‘true‘
else:
value = ‘false‘
elif isinstance(value, (list, tuple)):
value = u‘,‘.join([force_text(v) for v in value])
elif isinstance(value, (six.integer_types, float)):
# Leave it alone.
pass
else:
value = force_text(value)
return value
?
def _to_python(self, value):
"""
Converts values from Whoosh to native Python values.
?
A port of the same method in pysolr, as they deal with data the same way.
"""
if value == ‘true‘:
return True
elif value == ‘false‘:
return False
?
if value and isinstance(value, six.string_types):
possible_datetime = DATETIME_REGEX.search(value)
?
if possible_datetime:
date_values = possible_datetime.groupdict()
?
for dk, dv in date_values.items():
date_values[dk] = int(dv)
?
return datetime(date_values[‘year‘], date_values[‘month‘], date_values[‘day‘], date_values[‘hour‘],
date_values[‘minute‘], date_values[‘second‘])
?
try:
# Attempt to use json to load the values.
converted_value = json.loads(value)
?
# Try to handle most built-in types.
if isinstance(converted_value, (list, tuple, set, dict, six.integer_types, float, complex)):
return converted_value
except:
# If it fails (SyntaxError or its ilk) or we don‘t trust it,
# continue on.
pass
?
return value
?
?
class WhooshSearchQuery(BaseSearchQuery):
def _convert_datetime(self, date):
if hasattr(date, ‘hour‘):
return force_text(date.strftime(‘%Y%m%d%H%M%S‘))
else:
return force_text(date.strftime(‘%Y%m%d000000‘))
?
def clean(self, query_fragment):
"""
Provides a mechanism for sanitizing user input before presenting the
value to the backend.
?
Whoosh 1.X differs here in that you can no longer use a backslash
to escape reserved characters. Instead, the whole word should be
quoted.
"""
words = query_fragment.split()
cleaned_words = []
?
for word in words:
if word in self.backend.RESERVED_WORDS:
word = word.replace(word, word.lower())
?
for char in self.backend.RESERVED_CHARACTERS:
if char in word:
word = "‘%s‘" % word
break
?
cleaned_words.append(word)
?
return ‘ ‘.join(cleaned_words)
?
def build_query_fragment(self, field, filter_type, value):
from haystack import connections
query_frag = ‘‘
is_datetime = False
?
if not hasattr(value, ‘input_type_name‘):
# Handle when we‘ve got a ``ValuesListQuerySet``...
if hasattr(value, ‘values_list‘):
value = list(value)
?
if hasattr(value, ‘strftime‘):
is_datetime = True
?
if isinstance(value, six.string_types) and value != ‘ ‘:
# It‘s not an ``InputType``. Assume ``Clean``.
value = Clean(value)
else:
value = PythonData(value)
?
# Prepare the query using the InputType.
prepared_value = value.prepare(self)
?
if not isinstance(prepared_value, (set, list, tuple)):
# Then convert whatever we get back to what pysolr wants if needed.
prepared_value = self.backend._from_python(prepared_value)
?
# ‘content‘ is a special reserved word, much like ‘pk‘ in
# Django‘s ORM layer. It indicates ‘no special field‘.
if field == ‘content‘:
index_fieldname = ‘‘
else:
index_fieldname = u‘%s:‘ % connections[self._using].get_unified_index().get_index_fieldname(field)
?
filter_types = {
‘content‘: ‘%s‘,
‘contains‘: ‘*%s*‘,
‘endswith‘: "*%s",
‘startswith‘: "%s*",
‘exact‘: ‘%s‘,
‘gt‘: "{%s to}",
‘gte‘: "[%s to]",
‘lt‘: "{to %s}",
‘lte‘: "[to %s]",
‘fuzzy‘: u‘%s~‘,
}
?
if value.post_process is False:
query_frag = prepared_value
else:
if filter_type in [‘content‘, ‘contains‘, ‘startswith‘, ‘endswith‘, ‘fuzzy‘]:
if value.input_type_name == ‘exact‘:
query_frag = prepared_value
else:
# Iterate over terms & incorportate the converted form of each into the query.
terms = []
?
if isinstance(prepared_value, six.string_types):
possible_values = prepared_value.split(‘ ‘)
else:
if is_datetime is True:
prepared_value = self._convert_datetime(prepared_value)
?
possible_values = [prepared_value]
?
for possible_value in possible_values:
terms.append(filter_types[filter_type] % self.backend._from_python(possible_value))
?
if len(terms) == 1:
query_frag = terms[0]
else:
query_frag = u"(%s)" % " AND ".join(terms)
elif filter_type == ‘in‘:
in_options = []
?
for possible_value in prepared_value:
is_datetime = False
?
if hasattr(possible_value, ‘strftime‘):
is_datetime = True
?
pv = self.backend._from_python(possible_value)
?
if is_datetime is True:
pv = self._convert_datetime(pv)
?
if isinstance(pv, six.string_types) and not is_datetime:
in_options.append(‘"%s"‘ % pv)
else:
in_options.append(‘%s‘ % pv)
?
query_frag = "(%s)" % " OR ".join(in_options)
elif filter_type == ‘range‘:
start = self.backend._from_python(prepared_value[0])
end = self.backend._from_python(prepared_value[1])
?
if hasattr(prepared_value[0], ‘strftime‘):
start = self._convert_datetime(start)
?
if hasattr(prepared_value[1], ‘strftime‘):
end = self._convert_datetime(end)
?
query_frag = u"[%s to %s]" % (start, end)
elif filter_type == ‘exact‘:
if value.input_type_name == ‘exact‘:
query_frag = prepared_value
else:
prepared_value = Exact(prepared_value).prepare(self)
query_frag = filter_types[filter_type] % prepared_value
else:
if is_datetime is True:
prepared_value = self._convert_datetime(prepared_value)
?
query_frag = filter_types[filter_type] % prepared_value
?
if len(query_frag) and not isinstance(value, Raw):
if not query_frag.startswith(‘(‘) and not query_frag.endswith(‘)‘):
query_frag = "(%s)" % query_frag
?
return u"%s%s" % (index_fieldname, query_frag)
?
?
class WhooshEngine(BaseEngine):
backend = WhooshSearchBackend
query = WhooshSearchQuery
?
Linux入门Linux服务器搭建工作需要掌握的核心点
虚拟机的使用Linux安装(注意事项)服务器搭建(重点)网络配置(本地虚拟机)SSH连接远程服务器(putty、xshell6)FTP文件传输(FlashFXP、winscp)安装python(Linux自带python2.7.5)虚拟环境管理(virtualenv)django安装web服务器(Nginx + uwsgi) django项目发布数据库mysqlDNS解析(域名)Nginx多项目配置虚拟机安装虚拟机安装[重要]:https://blog.csdn.net/qq_39038465/article/details/81478847Linux目录结构bin:存放二进制可执行文件boot: 存放用于系统引导时使用的各种文件dev: 用于存放设备文件etc: 存放系统配置文件home: 存放所有用户文件的根目录lib 存放跟文件系统中的程序运行所需要的共享库及内核模块mnt 系统管理员安装 临时文件系统的安装点opt 额外安装 的可选应用程序包所放置的位置proc 虚拟文件系统,存放当前内存的映射root 超级用户目录sbin 存放二进制可执行文件,只有root才可以访问tmp 用于存放临时文件usr 用于存放系统应用程序var 用于存放运行时需要改变数据的文件Linux命令IP地址和主机名相关的命令查看IP:ifconfig重启网卡:service network restart查看网卡状态:service network status修改IP地址:vim /etc/sysconfig/network-scripts/ifcfg-ens33YPE="Ether net" # 网络类型,以太网BOOTPROTO="static"# 改为静态IPIPADDR="192.168.8.88"# IP地址NETMASK="255.255.255.0"# 子网掩码GATEWAY="192.168.8.1"# 网关DNS1="192.168.8.1"# 首选DNSONBOOT="yes"# 是否可以上网(默认为ON)1234567查看主机名:hostname修改主机名:vim /etc/hostnameCentOS7更换清华yum镜像清华:https://mirror.tuna.tsinghua.edu.cn/help/centos/
VIM常用命令 dd 删除光标所在的那一行u 撤销上一步操作ndd 删除光标所在位置起的多行 n为数字yy 复制光标当前所在的那一行nyy 复制多行 n为数字p 将已复制的内容粘贴到光标所在的位置的下一行大P 将已复制的内容粘贴到光标所在的位置的上一行np 粘贴多行到光标的下一行 n为数字ctrl+r 重复上一次操作$ 跳到一行的尾部0 跳到一行的头部gg 移动到这个文件的第一行G 跳到这个文件的最后一行nG 跳到n行set nu 显示行号H 光标移动到屏幕的最上方那一行的第一个字符M 光标移动到屏幕的中央那一行的第一个字符L 光标移动到屏幕的最下面那一行的第一个字符123456789101112131415161718CentOS7下安装python3# 1、安装pyhton3.7 的依赖包
yum -y groupinstall "Development tools"
yum -y install zlib-devel bzip2-devel openssl-devel ncurses-devel sqlite-devel readline-devel tk-devel gdbm-devel db4-devel libpcap-devel xz-devel libffi-devel
# 2、下载python3.7的“源码”: wget https://www.python.org/ftp/python/3.7.0/Python-3.7.0.tar.xz
# 3、解压并编译安装: tar -xJvf Python-3.7.0.tar.xz # 4、用cd命令进入解压出来的Python文件夹 cd Python-3.7.0
# 5、用./方法执行configure,并指定安装到usr目录下 ./configure --prefix=/usr/local/python3 --enable-shared
# 6、开始编译安装 make && make install
# 7、配置环境变量, 创建软链接 ln -s /usr/local/python3/bin/python3 /usr/bin/python3 # 创建python3的软链接 ln -s /usr/local/python3/bin/pip3 /usr/bin/pip3 # 创建pip的软链接# 8、将编译目录下的libpython3.7m.so.1.0文件复制到cp /root/Python-3.7.0/libpython3.7m.so.1.0 /usr/lib64/libpython3.7m.so.1.012345678910111213141516171819202122232425262728CentOS7下安装MySQL# 1、下载mysql的repo源 wget http://repo.mysql.com/mysql-community-release-el7-5.noarch.rpm
# 2、安装mysql-community-release-el7-5.noarch.rpm包 rpm -ivh mysql-community-release-el7-5.noarch.rpm
# 3、安装mysql yum install mysql-server
# 4、授权用户可以使用mysql chown -R root:root /var/lib/mysql
# 5、重启服务 service mysqld restart
# 6、接下来登录重置密码: mysql -u root # 进入mysql # 下面为mysql命令 use mysql; update user set password=password(‘root‘) where user=‘root‘; grant all privileges on *.* to ‘root‘@‘%‘ identified by ‘root‘; #设置远程登陆密码 flush privileges; #刷新当前配置 注:如果不管用,重启虚拟机ctrl+c,退出myql# 7、开放3306端口: # 设置 iptables serviceyum -y install iptables-services
# 如果要修改防火墙配置,如增加防火墙端口3306vi /etc/sysconfig/iptables
# 增加规则-A INPUT -p tcp -m state --state NEW -m tcp --dport 3306 -j ACCEPT #保存退出后
# 8、配置防火墙: systemctl restart iptables.service # 重启防火墙使配置生效 systemctl enable iptables.service # 设置防火墙开机启动12345678910111213141516171819202122232425262728293031323334353637连接MySQL数据库,并新建一个库以备的django使用使用navicat连接数据库
打开navicat以后,点击右上角的连接,弹出以下窗口,按下图中的内容填写
新建数据库,以便的django连接
数据库已经建好,下面就是django配置数据库django配置MySQL数据库# 1、在settings文件中,把数据库配置地方更改为以下内容:DATABASES = {‘default‘: { ‘ENGINE‘: ‘django.db.backends.mysql‘, # 默认数据库为MySQL ‘NAME‘: ‘library‘, # 数据库名为library ‘USER‘: ‘root‘, # 连接数据库的用户 "root" ‘PASSWORD‘: ‘123456‘, # 用户密码 "123456" ‘HOST‘: ‘www.XXXXXX.cn‘, # 主机的IP或者域名都可以 ‘PORT‘: 3306, # 数据库端口,默认为3306}}1234567891011django项目目录下的__init__.py文件中,导入pymysqlimport pymysqlpymysql.install_as_MySQLdb()12CentOS7安装虚拟环境# 安装虚拟环境pip3 install virtualenv
# 创建软链接ln -s /usr/local/python3/bin/virtualenv /usr/bin/virtualenv12345创建目录# 创建报错虚拟环境目录 名字是任意的mkdir -p /data/env # 个人网站发布文件夹 .名字都是任意的!mkdir -p /data/wwwroot1234创建、进入虚拟环境# 进入env目录cd /data/env# 创建虚拟环境virtualenv --python=/usr/bin/python3 py3_django2# 激活虚拟环境cd /data/env/py3_django2/binsource activate # 退出: deactivate# 安装django、uwsgi等.pip install djangopip install uwsgi # django项目发布相关# 退出虚拟环境cd /data/env/py3_django2/bindeactivate 12345678910111213为uwsgi创建软链接# 给uwsgi建立软链接,方便使用ln -s /usr/local/python3/bin/uwsgi /usr/bin/uwsgi12创建xml文件,保存名字与项目名同名,后缀为.xml<?xml version="1.0" encoding="UTF-8"?><uwsgi> <socket>127.0.0.1:8000</socket> <!-- 内部端口,自定义 --> <chdir>/data/wwwroot/library/</chdir> <!-- 项目路径 --> <module>library.wsgi</module> <!-- wsgi.py所在目录名--> <processes>4</processes> <!-- 进程数 --> <daemonize>uwsgi.log</daemonize> <!-- 日志文件 --></uwsgi>12345678本地项目上传到CentOS7服务器上# 1、在windows系统下,用cmd进入项目的目录,生成项目包依赖列表(如果依赖包少的话,这一步忽略)pip freeze > requirements.txt# 2、settings文件设置ALLOWED_HOSTS = [‘*‘] # 允许所有IP访问1234CentOS7下安装相关依赖包# CentOS7下安装相关依赖包(如果依赖包少的话,这一步忽略)pip install -r requirements.txt12迁移静态文件# 一、指定收集静态文件的目录,修改settings文件中静态文件路径 STATIC_ROOT = ‘/data/wwwroot/library/static‘
# 二、收集所有静态文件到STATIC_ROOT指定的目录python3 manage.py collectstatic12345安装Nginx# 1、用wget下载Nginxwget http://nginx.org/download/nginx-1.13.7.tar.gz # 2、下载完成后,解压tar -zxvf nginx-1.13.7.tar.gz
# 3、进入到nginx-1.13.7目录下,并执行以下命令./configuremake && make install
# 4、nginx一般默认安装好的路径为/usr/local/nginx 在/usr/local/nginx/conf/中先备份一下nginx.conf文件,以防意外。cd /usr/local/nginx/conf/cp nginx.conf nginx.conf.bak
# 5、然后打开nginx.conf,把原来的内容删除,直接加入以下内容:events { worker_connections 1024;}http { include mime.types; default_type application/octet-stream; sendfile on; server { listen 80; server_name www.xxxxxx.cn; #改为自己的域名,没域名修改为127.0.0.1:80 charset utf-8; location / { include uwsgi_params; uwsgi_pass 127.0.0.1:8000; #端口要和uwsgi里配置的一样 uwsgi_param UWSGI_SCRIPT library.wsgi; #wsgi.py所在的目录名+.wsgi uwsgi_param UWSGI_CHDIR /data/wwwroot/library/; #项目路径 } location /static/ { alias /data/wwwroot/library/static/; #静态资源路径 } }}
""" 6、 要留意备注的地方,要和UWSGI配置文件mysite.xml,还有项目路径对应上。 进入/usr/local/nginx/sbin/目录执行./nginx -t命令先检查配置文件是否有错,没有错就执行以下命令:"""cd /usr/local/nginx/sbin/./nginx
# 没有提示,证明成功# 测试127.0.0.1:80
12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152启动项目# 进入djnago项目cd /data/wwwroot/library/
# uwsgi 解析项目中的配置文件uwsgi -x library.xml
#以上步骤都没有出错的话。cd /usr/local/nginx/sbin/
# 重启nginx./nginx -s reload
# 服务器内部测试是否发布成功curl 1270.0.0.0:80 #就可以看到网站! # 关闭防火墙,否则远程不能访问!systemctl stop firewalld.service1234567891011121314151617Linux下防火墙管理# CentOS 7.0默认使用的是firewall作为防火墙,使用iptables必须重新设置一下
# 1、直接关闭防火墙systemctl stop firewalld.service # 停止firewallsystemctl disable firewalld.service # 禁止firewall开机启动
# 2、设置 iptables service (防火墙)yum -y install iptables-services # 安装防火墙管理# 如果要修改防火墙配置,如增加防火墙端口3306vi /etc/sysconfig/iptables # 用vi编辑器修改防火墙配置# 增加规则-A INPUT -p tcp -m state --state NEW -m tcp --dport 3306 -j ACCEPT# 保存退出后esc :wq systemctl restart iptables.service # 重启防火墙使配置生效systemctl enable iptables.service # 设置防火墙开机启动
123456789101112131415161718常用指令# 查看python版 本python -V
# 查看python命令如何解析which python3 # 找到命令位置 /usr/bin/python
# cd 命令进入到 /usr/binls -al python3*
# 查看虚拟环境目录和项目发布目录--------------------- 作者:董海明 来源:CSDN 原文:https://blog.csdn.net/haiming0415/article/details/89946928 版权声明:本文为博主原创文章,转载请附上博文链接!
标签:attr sso 证明 reverse padding return 主机名 failed elf
原文地址:https://www.cnblogs.com/guofeng-1016/p/11290685.html