package main import ( "database/sql" "fmt" _ "github.com/go-sql-driver/mysql" "html" "html/template" "io/ioutil" "log" "net/http" "net/url" "strconv" "strings" "unicode/utf8" // "sync" // "time" ) type indexPage struct{} type errorReport struct{ Error string } type surpriseURL struct{ Url string } type settingsPage struct{ Worksafe, FilterHTTPS bool } type MySQLResults struct{ Id, Url, Title, Description, Body string } type PageData struct { DBResults []MySQLResults Query, Totalcount string FindMore bool } func main() { http.HandleFunc("/", handler) http.HandleFunc("/json", handler) http.HandleFunc("/json/", handler) http.HandleFunc("/surprise", surprise) http.HandleFunc("/surprise/", surprise) http.HandleFunc("/settings/", settings) http.HandleFunc("/settings", settings) log.Fatal(http.ListenAndServe("0.0.0.0:8080", nil)) //set IP to 127.0.0.1 if reverse proxy is on the same machine } //https://golang.org/pkg/net/http/#Request func handler(w http.ResponseWriter, r *http.Request) { //fmt.Fprintf(w, "%s %s \n", r.Method, r.URL) //fmt.Fprintf(w, "%s \n", r.URL.RawQuery) //check if worksafe+https cookie enabled. filterHTTPS := false worksafe := true worksafeHTTPSCookie, err := r.Cookie("ws") if err != nil { worksafe = true filterHTTPS = false } else if worksafeHTTPSCookie.Value == "0" { worksafe = false filterHTTPS = false } else if worksafeHTTPSCookie.Value == "1" { worksafe = true filterHTTPS = false } else if worksafeHTTPSCookie.Value == "2" { worksafe = false filterHTTPS = true } else if worksafeHTTPSCookie.Value == "3" { worksafe = true filterHTTPS = true } //setup for error report error := errorReport{} //Get the raw query m, _ := url.ParseQuery(r.URL.RawQuery) //Get the query parameters (q and o) //fmt.Fprintf(w,"%s\n%s\n", m["q"][0], m["o"][0]) json := false if strings.Contains(r.URL.Path, "/json") { json = true if _, ok := m["nsfw"]; ok { //check if &nsfw added to json url worksafe = false } } query := "" queryNoQuotes := "" queryNoQuotes_SQLsafe := "" offset := "0" //Check if query and offset params exist if _, ok := m["q"]; ok { query = strings.Replace(m["q"][0], "'", "''", -1) queryNoQuotes = m["q"][0] } if _, ok := m["o"]; ok { offset = strings.Replace(m["o"][0], "'", "''", -1) } lim := "12" // limDistributedInt := if query == "" { //what do if no query found? //load index if no query detected if r.URL.Path == "/" { p := indexPage{} t, _ := template.ParseFiles("coreassets/form.html.go") t.Execute(w, p) } else if strings.Contains(r.URL.Path, "/json") { //load json info page if json selected p := indexPage{} t, _ := template.ParseFiles("coreassets/json/json.html.go") t.Execute(w, p) } else { p := indexPage{} t, _ := template.ParseFiles("coreassets/form.html.go") t.Execute(w, p) } } else { //Make sure offset is a number offsetInt, err := strconv.Atoi(offset) if err != nil { offset = "0" offsetInt = 0 } //Convert lim to number also limInt, _ := strconv.Atoi(lim) //get some details from the raw query var additions string querylen := len(query) //see if a search redirect (! or &) is used for a different search engine if json == false && (strings.Contains(m["q"][0],"!") || strings.Contains(m["q"][0],"&")){ searchredirect(w, r, m["q"][0]) } //phone users if query[querylen-1] == ' '{ query = query[:querylen-1] queryNoQuotes = queryNoQuotes[:len(queryNoQuotes)-1] querylen = len(query) } //check if user wants to limit search to a specific website sitePos := -1 siteEnd := 0 siteURL := "" if strings.Index(strings.ToLower(query), "site:") > -1 { //get url user wants to search and remove it from the query string sitePos = strings.Index(strings.ToLower(query), "site:") siteEnd = strings.Index(query[sitePos:], " ") //fmt.Printf("\n%d\n%d\n",sitePos,siteEnd) if siteEnd > -1 && sitePos > 1 { //site is not last part of query siteURL = query[sitePos+5 : siteEnd+sitePos] query = query[:sitePos-1] + query[siteEnd+sitePos:] queryNoQuotes = queryNoQuotes[:sitePos-1] + queryNoQuotes[siteEnd+sitePos:] additions = additions + "AND url LIKE '%" + siteURL + "%' " } else if siteEnd > -1 && sitePos == 0 { //site is at beginning siteURL = query[sitePos+5 : siteEnd] query = query[siteEnd+1:] queryNoQuotes = queryNoQuotes[siteEnd+1:] additions = additions + "AND url LIKE '%" + siteURL + "%' " } else if siteEnd < 0 && sitePos > 1 { //site is at end siteURL = query[sitePos+5:] query = query[:sitePos-1] queryNoQuotes = queryNoQuotes[:sitePos-1] additions = additions + "AND url LIKE '%" + siteURL + "%' " }else if querylen > 5{ query = query[5:] } querylen = len(query) } //fmt.Printf("Addition: \n%s\nQuery: '%s'\n",additions,query) //see if user uses -https flag (instead of cookie settings option) if querylen > 7 && strings.ToLower(query[querylen-7:querylen]) == " -https" { filterHTTPS = true query = query[0 : querylen-7] querylen = len(query) } //check if user wants to search within a time window (day,week,month) option := "" //fmt.Printf("\n'%s'\n",query) location := strings.Index(query, " !") if location == -1 { location = strings.Index(query, " &") } if location > -1 && strings.Index(query[location+1:querylen], " ") == -1 { //option is at end of query option = query[location+2 : querylen] query = query[:location] queryNoQuotes = queryNoQuotes[:location] querylen = len(query) }else if querylen > 0 && (query[0] == '!' || query[0] == '&') && strings.Index(query, " ") > -1{ //option is at start of query option = query[1:strings.Index(query, " ")] query = query[strings.Index(query, " ")+1:] queryNoQuotes = queryNoQuotes[strings.Index(queryNoQuotes, " ")+1:] querylen = len(query) } option = strings.ToLower(option) if option != "" { if option == "td" { //day additions = additions + "AND date > NOW() - INTERVAL 1 DAY " } else if option == "tw" { //week additions = additions + "AND date > NOW() - INTERVAL 7 DAY " } else if option == "tm" { //month additions = additions + "AND date > NOW() - INTERVAL 30 DAY " } else if option == "ty" { //year additions = additions + "AND date > NOW() - INTERVAL 365 DAY " } } //check if worksafe and filterHTTPS flags set if worksafe == true { additions = additions + "AND worksafe = '1' " } if filterHTTPS == true { additions = additions + "AND http = '1' " } //if query is just 1 or 2 letters, help make it work. Also CIA :D oneletterquery := 0 if len(query) < 3 || query == "cia" || query == "CIA" { queryfix := " " + query + " *" query = queryfix queryNoQuotes = queryfix if len(query) == 1 { oneletterquery = 1 } } //search if query has quotes and remove them (so we can find the longest word in the query) exactMatch := false //queryNoQuotes := query if strings.Contains(query, "\"") { exactMatch = true queryNoQuotes = strings.TrimLeft(queryNoQuotes, "\"") getlastquote := strings.Split(queryNoQuotes, "\"") queryNoQuotes = getlastquote[0] //fmt.Printf("%s \n", queryNoQuotes) } //Prepare to find longest word in query words := strings.Split(queryNoQuotes, " ") longestWordLength := 0 longestWord := "" wordcount := 0 longestwordelementnum := 0 queryNoQuotesOrFlags := "" requiredword := "" //queryNoFlags := "" //first remove any flags inside var queryNoQuotes, also grab any required words (+ prefix) if strings.Contains(queryNoQuotes, "-") || strings.Contains(queryNoQuotes, "+") { for i, wordNoFlags := range words { if i > 0 && strings.HasPrefix(wordNoFlags, "-") == false && strings.HasPrefix(wordNoFlags, "+") == false { queryNoQuotesOrFlags += " " } if strings.HasPrefix(wordNoFlags, "-") == false && strings.HasPrefix(wordNoFlags, "+") == false { queryNoQuotesOrFlags += wordNoFlags } if strings.HasPrefix(wordNoFlags, "+") == true && len(wordNoFlags) > 1 { //get requiredword requiredword = wordNoFlags[1:len(wordNoFlags)] } } queryNoQuotes = queryNoQuotesOrFlags } //now find longest word words = strings.Split(queryNoQuotes, " ") if exactMatch == false { for _, word := range words { if len(word) > longestWordLength { longestWordLength = len(word) longestWord = word longestwordelementnum = wordcount } wordcount++ } } //remove the '*' if contained anywhere in queryNoQuotes if strings.Contains(queryNoQuotes, "*") && exactMatch == false { queryNoQuotes = strings.Replace(queryNoQuotes, "*", "", -1) } //get sql safe querynoquotes queryNoQuotes_SQLsafe = strings.Replace(queryNoQuotes, "'", "''", -1) //fmt.Fprintf(w,"%s\n%s\n", query,offset) //fmt.Printf("hai\n") //get copy of original query because we might have to modify it somewhat queryOriginal := query tRes := MySQLResults{} var res = PageData{} //Check if query is a url urlDetected := false isURL := "" if strings.Index(query, " ") == -1 && strings.Index(query, "\"") == -1 && strings.Index(query, ".") > -1 { //note this will also flag on file extensions if len(query) > 6 && (query[0:7] == "http://" || query[0:7] == "HTTP://") { query = query[7:] } else if len(query) > 7 && (query[0:8] == "https://" || query[0:8] == "HTTPS://") { query = query[8:] } if len(queryNoQuotes_SQLsafe) > 6 && (queryNoQuotes_SQLsafe[0:7] == "http://" || queryNoQuotes_SQLsafe[0:7] == "HTTP://") { queryNoQuotes_SQLsafe = queryNoQuotes_SQLsafe[7:] } else if len(queryNoQuotes_SQLsafe) > 7 && (queryNoQuotes_SQLsafe[0:8] == "https://" || queryNoQuotes_SQLsafe[0:8] == "HTTPS://") { queryNoQuotes_SQLsafe = queryNoQuotes_SQLsafe[8:] } query = "\"" + query + "\"" urlDetected = true isURL = "WHEN LOCATE('" + queryNoQuotes_SQLsafe + "',url)>0 THEN 25" } //Check if query contains a hyphenated word. Will wrap quotes around hyphenated words that aren't part of a string which is already wraped in quotes. if (strings.Contains(queryNoQuotes_SQLsafe, "-") || strings.Contains(queryNoQuotes_SQLsafe, "+")) && urlDetected == false { if query == "c++" || query == "C++" { //shitty but works for now query = "c++ programming" } hyphenwords := strings.Split(query, " ") query = "" quotes := 0 for i, word := range hyphenwords { if strings.Contains(word, "\"") { quotes++ } if ((strings.Contains(word, "-") && word[0] != '-') || (strings.Contains(word, "+") && word[0] != '+')) && quotes%2 == 0 { //if hyphen or plus exists, not a flag, not wrapped in quotes already word = "\"" + word + "\"" } if i > 0 { query += " " } query += word } } //fmt.Printf(">%s<\n", query) //perform full text search FOR InnoDB or MyISAM var sqlQuery, id, url, title, description, body, idList string rangeOffset := 0 serverCount := 0 var servers []string numServers := 0 //parse res.csv noservers := 0 repLim, _ := strconv.Atoi(lim) repOffset, _ := strconv.Atoi(offset) repLimStr := "" repOffsetStr := "" noresults := 0 repsearchfail := 0 var idListChans []chan string resourceFile, err := ioutil.ReadFile("res.csv") if err != nil { noservers = 1 } else { if len(resourceFile) < 2 { noservers = 1 } } //this switches off use of concurrent slaves to process a one word query. Should remove this if the database grows significantly larger if strings.Contains(query, " ") == false && oneletterquery == 0 { noservers = 1 } if noservers == 0 { //send query to go routines. resourceFilestring := string(resourceFile) //just in case user is messing around res.csv with a text editor and the editor ads a line feed to the end of the file if len(resourceFilestring) > 0 && resourceFilestring[len(resourceFilestring)-1] == byte('\n') { resourceFilestring = resourceFilestring[0 : len(resourceFilestring)-1] } servers = strings.Split(resourceFilestring, "\n") numServers = len(servers) //numServers must divide evenly into lim, or lim must divide evenly into numservers //if they do not, automatically adjust numServers until they divide evenly //calculate number of servers to use based on lim size if limInt > numServers { for limInt%numServers > 0 { numServers -= 1 } } else if numServers > limInt { for numServers%limInt > 0 { numServers -= 1 } } //calculate limit and offset on distributed servers. if numServers < limInt { repLim = limInt / numServers } else { repLim = 1 } repOffset = offsetInt / numServers //calculate rangeOffset (offset for the range of returned results, important if numServers > 2*lim) rangeOffset = offsetInt - (repOffset * numServers) repLimStr = strconv.Itoa(repLim) repOffsetStr = strconv.Itoa(repOffset) //create a channel for each available server for i := 0; i < numServers; i++ { idListChans = append(idListChans, make(chan string)) } for _, server := range servers { serverSettings := strings.Split(server, ",") if len(serverSettings) == 4 { //if line contains all 4 settings //ip, database, startID, endID //create SQL connection string //db, err := sql.Open("mysql", "remote_guest:d0gemuchw0w@tcp(10.8.0.102:3306)/wiby?charset=utf8mb4") serverIP := serverSettings[0] serverDB := serverSettings[1] startID := serverSettings[2] endID := serverSettings[3] sqlString := "remote_guest:d0gemuchw0w@tcp(" + serverIP + ":3306)/" + serverDB + "?charset=utf8mb4" // fmt.Printf("%s %s %s %d\n",sqlString,startID,endID,numServers) //send special distributed query, only need ID returned and also search between startID/endID sqlQuery = "SELECT id FROM windex WHERE Match(tags, body, description, title, url) Against('" + query + "' IN BOOLEAN MODE) AND id BETWEEN " + startID + " AND " + endID + " AND enable = '1' " + additions + "ORDER BY CASE WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', tags)>0 THEN 30 " + isURL + " WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', title)>0 AND Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) THEN 20 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', title)>0 THEN 16 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', body)>0 THEN 15 WHEN Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) THEN Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) END DESC LIMIT " + repLimStr + " OFFSET " + repOffsetStr + "" go distributedQuery(sqlString, sqlQuery, startID, endID, idListChans[serverCount]) serverCount++ } } for i := 0; i < serverCount; i++ { //wait for channels to complete and collect results idList += <-idListChans[i] } if len(idList) > 0 { switch strings.Contains(idList, "e") { case true: repsearchfail = 1 default: idList = idList[1:len(idList)] //trim the first comma in the list } } else { noresults = 1 } //fmt.Printf("\nChan: %s",idList) } //if all went well with replication servers, send query to master containing idList and use the rangeOffset if numServers == serverCount && numServers > 0 && repsearchfail == 0 { sqlQuery = "SELECT id, url, title, description, body FROM windex WHERE Match(tags, body, description, title, url) Against('" + query + "' IN BOOLEAN MODE) AND id IN (" + idList + ") AND enable = '1' " + additions + "ORDER BY CASE WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', tags)>0 THEN 30 " + isURL + " WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', title)>0 AND Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) THEN 20 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', title)>0 THEN 16 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', body)>0 THEN 15 WHEN Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) THEN Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) END DESC LIMIT " + lim + " OFFSET " + strconv.Itoa(rangeOffset) + "" } else { //else, if no replication servers or there was some sort of error, just search the database locally instead sqlQuery = "SELECT id, url, title, description, body FROM windex WHERE Match(tags, body, description, title, url) Against('" + query + "' IN BOOLEAN MODE) AND enable = '1' " + additions + "ORDER BY CASE WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', tags)>0 THEN 30 " + isURL + " WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', title)>0 AND Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) THEN 20 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', title)>0 THEN 16 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', body)>0 THEN 15 WHEN Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) THEN Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) END DESC LIMIT " + lim + " OFFSET " + offset + "" } //init the db and set charset //create SQL connection string db, err := sql.Open("mysql", "guest:qwer@/wiby?charset=utf8mb4") if err != nil { p := indexPage{} t, _ := template.ParseFiles("coreassets/error.html.go") t.Execute(w, p) } defer db.Close() // If Open doesn't open a connection. Validate DSN data: err = db.Ping() if err != nil { error.Error = err.Error() t, _ := template.ParseFiles("coreassets/error.html.go") t.Execute(w, error) } count := 0 countResults := 0 switch noresults { //if noresults == 1, no results were found during search on active replication servers case 0: // Send the query rows, err := db.Query(sqlQuery) if err != nil { fmt.Printf("\n%s", err) res.Totalcount = strconv.Itoa(0) res.Query = m["q"][0] //get original unsafe query if json { w.Header().Set("Content-Type", "application/json") t, _ := template.ParseFiles("coreassets/json/results.json.go") t.Execute(w, res) } else { t, _ := template.ParseFiles("coreassets/results.html.go") t.Execute(w, res) } //p := indexPage{} //t, _ := template.ParseFiles("coreassets/form.html.go") //t.Execute(w, p) return } if urlDetected == true { query = queryOriginal } for rows.Next() { count++ countResults++ //this will get set if position of longest word of query is found within body pos := -1 err := rows.Scan(&id, &url, &title, &description, &body) if err != nil { error.Error = err.Error() t, _ := template.ParseFiles("coreassets/error.html.go") t.Execute(w, error) } //find query inside body of page if exactMatch == false { //remove the '*' if contained anywhere in query /*if strings.Contains(queryNoQuotes,"*"){ queryNoQuotes = strings.Replace(queryNoQuotes, "*", "", -1) } */ if len(requiredword) > 0 { //search for position of required word if any, else search for position of whole query pos = strings.Index(strings.ToLower(body), strings.ToLower(requiredword)) } else if pos == -1 { pos = strings.Index(strings.ToLower(body), strings.ToLower(queryNoQuotes)) } if pos == -1 { //prepare to find position of longest query word (or required word) within body //remove the '*' at the end of the longest word if present if strings.Contains(longestWord, "*") { longestWord = strings.Replace(longestWord, "*", "", -1) } //search within body for position of longest query word. pos = strings.Index(strings.ToLower(body), strings.ToLower(longestWord)) //not found?, set position to a different word, make sure there's no wildcard on it if pos == -1 && wordcount > 1 { if longestwordelementnum > 0 { words[0] = strings.Replace(words[0], "*", "", -1) pos = strings.Index(strings.ToLower(body), strings.ToLower(words[0])) } if longestwordelementnum == 0 { words[1] = strings.Replace(words[1], "*", "", -1) pos = strings.Index(strings.ToLower(body), strings.ToLower(words[1])) } } } } else { //if exact match, find position of query within body pos = strings.Index(strings.ToLower(body), strings.ToLower(queryNoQuotes)) } //still not found?, set position to 0 if pos == -1 { pos = 0 } //Adjust position for runes within body pos = utf8.RuneCountInString(body[:pos]) starttext := 0 //ballpark := 0 ballparktext := "" //figure out how much preceding text to use if pos < 32 { starttext = 0 } else if pos > 25 { starttext = pos - 25 } else if pos > 20 { starttext = pos - 15 } //total length of the ballpark textlength := 180 //populate the ballpark if pos >= 0 { ballparktext = substr(body, starttext, starttext+textlength) } //else{ ballpark = 0}//looks unused //find position of nearest Period //foundPeriod := true posPeriod := strings.Index(ballparktext, ". ") + starttext + 1 //find position of nearest Space //foundSpace := true posSpace := strings.Index(ballparktext, " ") + starttext //if longest word in query is after a period+space within ballpark, reset starttext to that point if (pos - starttext) > posPeriod { starttext = posPeriod //populate the bodymatch if (pos - starttext) >= 0 { body = substr(body, starttext, starttext+textlength) } else { body = "" } } else if pos > posSpace { //else if longest word in query is after a space within ballpark, reset starttext to that point //else if(pos-starttext) > posSpace//else if longest word in query is after a space within ballpark, reset starttext to that point starttext = posSpace //populate the bodymatch if (pos - starttext) >= 0 { body = substr(body, starttext, starttext+textlength) } else { body = "" } } else //else just set the bodymatch to the ballparktext { //populate the bodymatch if (pos - starttext) >= 0 { body = ballparktext } else { body = "" } } tRes.Id = id tRes.Url = url tRes.Title = html.UnescapeString(title) tRes.Description = html.UnescapeString(description) tRes.Body = html.UnescapeString(body) if json == true { tRes.Title = JSONRealEscapeString(tRes.Title) tRes.Description = JSONRealEscapeString(tRes.Description) tRes.Body = JSONRealEscapeString(tRes.Body) } res.DBResults = append(res.DBResults, tRes) } defer rows.Close() rows.Close() if count > 0 { //new search method may cause less than the limit of row results per page even if there are more results to come, so we force a full count count = limInt } } //end switch //================================================================================================================================ //no results found (count==0), so do a wildcard search (repeat the above process) addWildcard := false if count == 0 && offset == "0" && urlDetected == false && exactMatch == false { addWildcard = true query = strings.Replace(query, "\"", "", -1) //remove some things innodb gets fussy over query = strings.Replace(query, "*", "", -1) query = strings.Replace(query, "'", "", -1) queryNoQuotes_SQLsafe = strings.Replace(queryNoQuotes_SQLsafe, "\"", "", -1) queryNoQuotes_SQLsafe = strings.Replace(queryNoQuotes_SQLsafe, "*", "", -1) queryNoQuotes_SQLsafe = strings.Replace(queryNoQuotes_SQLsafe, "'", "", -1) query = query + "*" sqlQuery = "SELECT id, url, title, description, body FROM windex WHERE Match(tags, body, description, title, url) Against('" + query + "' IN BOOLEAN MODE) AND enable = '1' " + additions + "ORDER BY CASE WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', tags)>0 THEN 30 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', title)>0 AND Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) THEN 20 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', title)>0 THEN 16 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', body)>0 THEN 15 WHEN Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) THEN Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) END DESC LIMIT " + lim + " OFFSET " + offset + "" if repsearchfail == 0 && noservers == 0 { serverCount = 0 idList = "" for _, server := range servers { serverSettings := strings.Split(server, ",") if len(serverSettings) == 4 { //if line contains all 4 settings //ip, database, startID, endID //create SQL connection string //db, err := sql.Open("mysql", "remote_guest:d0gemuchw0w@tcp(10.8.0.102:3306)/wiby?charset=utf8mb4") serverIP := serverSettings[0] serverDB := serverSettings[1] startID := serverSettings[2] endID := serverSettings[3] sqlString := "remote_guest:d0gemuchw0w@tcp(" + serverIP + ":3306)/" + serverDB + "?charset=utf8mb4" //fmt.Printf("%s %s %s %d\n",sqlString,startID,endID,numServers) //send special distributed query, only need ID returned and also search between startID/endID sqlQuery = "SELECT id FROM windex WHERE Match(tags, body, description, title, url) Against('" + query + "' IN BOOLEAN MODE) AND id BETWEEN " + startID + " AND " + endID + " AND enable = '1' " + additions + "ORDER BY CASE WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', tags)>0 THEN 30 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', title)>0 AND Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) THEN 20 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', title)>0 THEN 16 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', body)>0 THEN 15 WHEN Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) THEN Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) END DESC LIMIT " + repLimStr + " OFFSET " + repOffsetStr + "" go distributedQuery(sqlString, sqlQuery, startID, endID, idListChans[serverCount]) serverCount++ } } for i := 0; i < serverCount; i++ { //wait for channels to complete and collect results idList += <-idListChans[i] } if len(idList) > 0 { switch strings.Contains(idList, "e") { case true: repsearchfail = 1 default: idList = idList[1:len(idList)] //trim the first comma in the list } } else { noresults = 1 } //if all went well with replication servers, send query to master containing idList and use the rangeOffset if numServers == serverCount && numServers > 0 && repsearchfail == 0 { sqlQuery = "SELECT id, url, title, description, body FROM windex WHERE Match(tags, body, description, title, url) Against('" + query + "' IN BOOLEAN MODE) AND id IN (" + idList + ") AND enable = '1' " + additions + "ORDER BY CASE WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', tags)>0 THEN 30 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', title)>0 AND Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) THEN 20 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', title)>0 THEN 16 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', body)>0 THEN 15 WHEN Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) THEN Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) END DESC LIMIT " + lim + " OFFSET " + strconv.Itoa(rangeOffset) + "" } else { //else, if no replication servers or there was some sort of error, just search the database locally instead sqlQuery = "SELECT id, url, title, description, body FROM windex WHERE Match(tags, body, description, title, url) Against('" + query + "' IN BOOLEAN MODE) AND enable = '1' " + additions + "ORDER BY CASE WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', tags)>0 THEN 30 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', title)>0 AND Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) THEN 20 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', title)>0 THEN 16 WHEN LOCATE('" + queryNoQuotes_SQLsafe + "', body)>0 THEN 15 WHEN Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) THEN Match(title) AGAINST('" + query + "' IN BOOLEAN MODE) END DESC LIMIT " + lim + " OFFSET " + offset + "" } } rows2, err := db.Query(sqlQuery) if err != nil { res.Totalcount = strconv.Itoa(0) res.Query = m["q"][0] //get original unsafe query if json { w.Header().Set("Content-Type", "application/json") t, _ := template.ParseFiles("coreassets/json/results.json.go") t.Execute(w, res) } else { t, _ := template.ParseFiles("coreassets/results.html.go") t.Execute(w, res) } return } for rows2.Next() { count++ //this will get set if position of longest word of query is found within body pos := -1 err := rows2.Scan(&id, &url, &title, &description, &body) if err != nil { error.Error = err.Error() t, _ := template.ParseFiles("coreassets/error.html.go") t.Execute(w, error) } //find query inside body of page if exactMatch == false { if len(requiredword) > 0 { //search for position of required word if any, else search for position of whole query pos = strings.Index(strings.ToLower(body), strings.ToLower(requiredword)) } else if pos == -1 { pos = strings.Index(strings.ToLower(body), strings.ToLower(queryNoQuotes)) } if pos == -1 { //prepare to find position of longest query word (or required word) within body //remove the '*' at the end of the longest word if present if strings.Contains(longestWord, "*") { longestWord = strings.Replace(longestWord, "*", "", -1) } //search within body for position of longest query word. pos = strings.Index(strings.ToLower(body), strings.ToLower(longestWord)) //not found?, set position to a different word, make sure there's no wildcard on it if pos == -1 && wordcount > 1 { if longestwordelementnum > 0 { words[0] = strings.Replace(words[0], "*", "", -1) pos = strings.Index(strings.ToLower(body), strings.ToLower(words[0])) } if longestwordelementnum == 0 { words[1] = strings.Replace(words[1], "*", "", -1) pos = strings.Index(strings.ToLower(body), strings.ToLower(words[1])) } } } } else { //if exact match, find position of query within body pos = strings.Index(strings.ToLower(body), strings.ToLower(queryNoQuotes)) } //still not found?, set position to 0 if pos == -1 { pos = 0 } //Adjust position for runes within body pos = utf8.RuneCountInString(body[:pos]) starttext := 0 //ballpark := 0 ballparktext := "" //figure out how much preceding text to use if pos < 32 { starttext = 0 } else if pos > 25 { starttext = pos - 25 } else if pos > 20 { starttext = pos - 15 } //total length of the ballpark textlength := 180 //populate the ballpark if pos >= 0 { ballparktext = substr(body, starttext, starttext+textlength) } //else{ ballpark = 0}//looks unused //find position of nearest Period //foundPeriod := true posPeriod := strings.Index(ballparktext, ". ") + starttext + 1 //find position of nearest Space //foundSpace := true posSpace := strings.Index(ballparktext, " ") + starttext //if longest word in query is after a period+space within ballpark, reset starttext to that point if (pos - starttext) > posPeriod { starttext = posPeriod //populate the bodymatch if (pos - starttext) >= 0 { body = substr(body, starttext, starttext+textlength) } else { body = "" } } else if pos > posSpace { //else if longest word in query is after a space within ballpark, reset starttext to that point //else if(pos-starttext) > posSpace//else if longest word in query is after a space within ballpark, reset starttext to that point starttext = posSpace //populate the bodymatch if (pos - starttext) >= 0 { body = substr(body, starttext, starttext+textlength) } else { body = "" } } else //else just set the bodymatch to the ballparktext { //populate the bodymatch if (pos - starttext) >= 0 { body = ballparktext } else { body = "" } } tRes.Id = id tRes.Url = url tRes.Title = html.UnescapeString(title) tRes.Description = html.UnescapeString(description) tRes.Body = html.UnescapeString(body) if json == true { tRes.Title = JSONRealEscapeString(tRes.Title) tRes.Description = JSONRealEscapeString(tRes.Description) tRes.Body = JSONRealEscapeString(tRes.Body) } res.DBResults = append(res.DBResults, tRes) } defer rows2.Close() rows2.Close() } //======================================================================================================================= //http://go-database-sql.org/retrieving.html //Close DB db.Close() //Allow the find more link if (countResults >= limInt && addWildcard == false) || addWildcard == false && countResults > 2 { res.FindMore = true } else { res.FindMore = false } totalCountInt := count + offsetInt res.Totalcount = strconv.Itoa(totalCountInt) res.Query = m["q"][0] //get original unsafe query if json { w.Header().Set("Content-Type", "application/json") t, _ := template.ParseFiles("coreassets/json/results.json.go") t.Execute(w, res) } else { t, _ := template.ParseFiles("coreassets/results.html.go") t.Execute(w, res) } } } func settings(w http.ResponseWriter, r *http.Request) { //setup for error report error := errorReport{} //check if worksafe (adult content) cookie enabled. filterHTTPS := false worksafe := true worksafewasoff := false worksafeHTTPSCookie, err := r.Cookie("ws") if err != nil { worksafe = true filterHTTPS = false } else if worksafeHTTPSCookie.Value == "0" { worksafe = false filterHTTPS = false worksafewasoff = true } else if worksafeHTTPSCookie.Value == "1" { worksafe = true filterHTTPS = false } else if worksafeHTTPSCookie.Value == "2" { worksafe = false filterHTTPS = true worksafewasoff = true } else if worksafeHTTPSCookie.Value == "3" { worksafe = true filterHTTPS = true } //check if and what is the user posting switch r.Method { case "POST": if err := r.ParseForm(); err != nil { error.Error = err.Error() t, _ := template.ParseFiles("coreassets/error.html.go") t.Execute(w, error) } worksafebox := r.Form.Get("worksafe") agreecheck := r.Form.Get("agree") agreesubmit := r.Form.Get("agreesubmit") httpsbox := r.Form.Get("filterHTTPS") //if user agrees to terms to disable adult content, set cookie and return to index if agreecheck == "on" { worksafe = false //expiration := time.Now().Add(365 * 24 * time.Hour) if filterHTTPS == false { cookie := http.Cookie{Name: "ws", Value: "0", Path: "/"} http.SetCookie(w, &cookie) } else { cookie := http.Cookie{Name: "ws", Value: "2", Path: "/"} http.SetCookie(w, &cookie) } p := indexPage{} t, _ := template.ParseFiles("coreassets/settings/gohome.html") t.Execute(w, p) //else if worksafebox is checked, return to index with worksafe on } else if worksafebox == "on" || agreesubmit == "on" { //expiration := time.Now().Add(365 * 24 * time.Hour) if httpsbox != "on" { cookie := http.Cookie{Name: "ws", Value: "1", Path: "/"} http.SetCookie(w, &cookie) } else { cookie := http.Cookie{Name: "ws", Value: "3", Path: "/"} http.SetCookie(w, &cookie) } p := indexPage{} t, _ := template.ParseFiles("coreassets/settings/gohome.html") t.Execute(w, p) //else if worksafebox unchecked and no cookie, go to content agreement section } else if worksafebox != "on" && worksafewasoff == false && agreesubmit != "on" { p := indexPage{} if httpsbox == "on" { cookie := http.Cookie{Name: "ws", Value: "3", Path: "/"} http.SetCookie(w, &cookie) } else { cookie := http.Cookie{Name: "ws", Value: "1", Path: "/"} http.SetCookie(w, &cookie) } t, _ := template.ParseFiles("coreassets/settings/agree.html.go") t.Execute(w, p) //else if worksafebox unchecked and cookie alredy agreed, go back to index } else if worksafebox != "on" && worksafewasoff == true { if httpsbox == "on" { cookie := http.Cookie{Name: "ws", Value: "2", Path: "/"} http.SetCookie(w, &cookie) } else { cookie := http.Cookie{Name: "ws", Value: "0", Path: "/"} http.SetCookie(w, &cookie) } p := indexPage{} t, _ := template.ParseFiles("coreassets/settings/gohome.html") t.Execute(w, p) } default: //load the settings page if no post value settingspage := settingsPage{} settingspage.Worksafe = worksafe settingspage.FilterHTTPS = filterHTTPS t, _ := template.ParseFiles("coreassets/settings/settings.html.go") t.Execute(w, settingspage) } } func surprise(w http.ResponseWriter, r *http.Request) { surprise := surpriseURL{} //check if worksafe+HTTPS cookie enabled. filterHTTPS := false worksafeHTTPSCookie, err := r.Cookie("ws") if err != nil { filterHTTPS = false } else if worksafeHTTPSCookie.Value == "2" { filterHTTPS = true } else if worksafeHTTPSCookie.Value == "3" { filterHTTPS = true } //setup for error report error := errorReport{} //init the db and set charset db, err := sql.Open("mysql", "guest:qwer@/wiby?charset=utf8mb4") if err != nil { error.Error = err.Error() t, _ := template.ParseFiles("coreassets/error.html.go") t.Execute(w, error) } defer db.Close() // Open doesn't open a connection. Validate DSN data: err = db.Ping() if err != nil { error.Error = err.Error() t, _ := template.ParseFiles("coreassets/error.html.go") t.Execute(w, error) } //grab a random page var sqlQuery string if filterHTTPS == false { sqlQuery = "select url from windex where worksafe = 1 and surprise = 1 order by rand() limit 1" } else { sqlQuery = "select url from windex where worksafe = 1 and surprise = 1 and http = 1 order by rand() limit 1" } rows, err := db.Query(sqlQuery) if err != nil { error.Error = err.Error() t, _ := template.ParseFiles("coreassets/error.html.go") t.Execute(w, error) } var url string for rows.Next() { err := rows.Scan(&url) if err != nil { error.Error = err.Error() t, _ := template.ParseFiles("coreassets/error.html.go") t.Execute(w, error) } surprise.Url = url } defer rows.Close() rows.Close() db.Close() t, _ := template.ParseFiles("coreassets/surprise.html.go") t.Execute(w, surprise) } func MysqlRealEscapeString(value string) string { replace := map[string]string{"\\": "\\\\", "'": `\'`, "\\0": "\\\\0", "\n": "\\n", "\r": "\\r", `"`: `\"`, "\x1a": "\\Z"} for b, a := range replace { value = strings.Replace(value, b, a, -1) } return value } func JSONRealEscapeString(value string) string { replace := map[string]string{"\\": "\\\\", "\t": "\\t", "\b": "\\b", "\n": "\\n", "\r": "\\r", "\f": "\\f" /*, `"`:`\"`*/} for b, a := range replace { value = strings.Replace(value, b, a, -1) } return value } func substr(s string, start int, end int) string { start_str_idx := 0 i := 0 for j := range s { if i == start { start_str_idx = j } if i == end { return s[start_str_idx:j] } i++ } return s[start_str_idx:] } func searchredirect(w http.ResponseWriter, r *http.Request, query string) { //separate actual query from search redirect actualquery := "" redirect := "" lenquery := len(query) if strings.Index(query," ") > -1{ location := strings.Index(query, " !") if location == -1 { location = strings.Index(query, " &") } if location > -1 && strings.Index(query[location+1:lenquery], " ") == -1 { //redirect is at end of query redirect = query[location+2 : lenquery] actualquery = query[:location] } else if (strings.Index(query, "!") == 0 || strings.Index(query, "&") == 0){ //redirect is at start of query redirect = query[1:strings.Index(query, " ")] actualquery = query[strings.Index(query, " ")+1:] //fmt.Printf("\nRedirect: %s\nquery: %s\n",redirect,actualquery) } redirect = strings.ToLower(redirect) }else if (query[0] == '!' || query[0] == '&') && lenquery > 1{ redirect = query[1:] } if redirect != "" { //determine which search engine to redirect if redirect == "g" { //if google text search http.Redirect(w, r, "http://google.com/search?q="+actualquery, http.StatusSeeOther) } else if redirect == "b" { //if bing text search http.Redirect(w, r, "http://bing.com/search?q="+actualquery, http.StatusSeeOther) } else if redirect == "gi" { //if google image search http.Redirect(w, r, "http://www.google.com/search?tbm=isch&q="+actualquery, http.StatusSeeOther) } else if redirect == "bi" { //if bing image search http.Redirect(w, r, "http://www.bing.com/images/search?q="+actualquery, http.StatusSeeOther) } else if redirect == "gv" { //if google video search http.Redirect(w, r, "http://www.google.com/search?tbm=vid&q="+actualquery, http.StatusSeeOther) } else if redirect == "bv" { //if bing video search http.Redirect(w, r, "http://www.bing.com/videos/search?q="+actualquery, http.StatusSeeOther) } else if redirect == "gm" { //if google maps search http.Redirect(w, r, "http://www.google.com/maps/search/"+actualquery, http.StatusSeeOther) } else if redirect == "bm" { //if bing maps search http.Redirect(w, r, "http://www.bing.com/maps?q="+actualquery, http.StatusSeeOther) }/* else { http.Redirect(w, r, "/?q="+actualquery, http.StatusSeeOther) }*/ } } func distributedQuery(con string, sqlQuery string, startID string, endID string, idListChan chan<- string) { var id string var idList string count := 0 //defer wg.Done() //init the db db, err := sql.Open("mysql", con) if err != nil { idList = idList + "e" //will look for this when channels are processed } defer db.Close() // If Open doesn't open a connection. Validate DSN data: err = db.Ping() if err != nil { } // Send the query rows, err := db.Query(sqlQuery) if err == nil { for rows.Next() { err := rows.Scan(&id) if err != nil { } //idString = idstring + "id = " + id + " or " idList += "," + id count++ } } else { idList = idList + "e" //will look for this when channels are processed fmt.Printf("%s", err) } //fmt.Printf("%s\n", idList) idListChan <- idList } /*func caseInsenstiveContains(fullstring, substring string) bool { return strings.Contains(strings.ToLower(fullstring), strings.ToLower(substring)) }*/ /* A QueryString is, by definition, in the URL. You can access the URL of the request using req.URL (doc). The URL object has a Query() method (doc) that returns a Values type, which is simply a map[string][]string of the QueryString parameters. If what you're looking for is the POST data as submitted by an HTML form, then this is (usually) a key-value pair in the request body. You're correct in your answer that you can call ParseForm() and then use req.Form field to get the map of key-value pairs, but you can also call FormValue(key) to get the value of a specific key. This calls ParseForm() if required, and gets values regardless of how they were sent (i.e. in query string or in the request body). req.URL.RawQuery returns everything after the ? on a GET request, if that helps. */ /*import ( "net/http" ) func main() { http.Handle("/", http.StripPrefix("/", http.FileServer(http.Dir("./")))) if err := http.ListenAndServe(":8080", nil); err != nil { panic(err) } }*/ /*func handler(w http.ResponseWriter, r *http.Request) { fmt.Fprintf(w, "%s %s %s \n", r.Method, r.URL, r.Proto) //Iterate over all header fields for k, v := range r.Header { fmt.Fprintf(w, "Header field %q, Value %q\n", k, v) } fmt.Fprintf(w, "Host = %q\n", r.Host) fmt.Fprintf(w, "RemoteAddr= %q\n", r.RemoteAddr) //Get value for a specified token fmt.Fprintf(w, "\n\nFinding value of \"Accept\" %q", r.Header["Accept"]) }*/