Sublime Forum

[SOLVED] Possible to launch custom AutoComplete on_modified OR include word separators on_query_completions?

#2

I’m not entirely sure I’m understanding… When the user types R, it seems that at the moment, you are showing the auto complete, with entries starting R., am I right? But you want to just show the R-related auto complete items when they type R. instead and drop the R. prefix from the entries? Would it not help to just trigger the autocomplete on ., check what comes before it, and if it is R, show the relevant entries?

Can you share some code, it might help me to understand better? :slightly_smiling:

side note: maybe a scope filter would be better than a file extension filter?

1 Like

#3

 
That’s pretty much what I’m trying to accomplish, and was attempting to describe in the OP. 

The second GIF demonstrates the process that would be used to verify the preceding word.

 

The problem is that . is a word_separator, which:

  • does not call on_query_completions
  • closes the auto-complete panel ( rather than than trigger it )

so I’m unable to implement the code that would verify the preceding word of a . character.
 



 

 

Here’s the on_query_completions code:
 

class EventListener ( sublime_plugin.EventListener ):

	def on_query_completions ( self, view, prefix, locations ):

		#▒▒▒  Get JSON Data  ▒▒▒#

		customCompletion_Files = sublime.find_resources ( "*.custom-completions" )
		file = customCompletion_Files[0]

		jsonString = sublime.load_resource ( file )
		jsonValues = sublime.decode_value ( jsonString )

		fileExtensions      = jsonValues[ "fileExtensions" ]
		completionSeparator = jsonValues[ "separator" ]
		jsonCompletions     = jsonValues[ "completions" ]

		if len ( fileExtensions ) == 0:
			fileExtension_Verification_Enabled = False
		else:
			fileExtension_Verification_Enabled = True

		#▒▒▒  Verify File Extension  ▒▒▒#

		fileExtension_Match_Found = False

		for extension in fileExtensions:
			if extension == V.fileExtension:
				fileExtension_Match_Found = True;

		if  fileExtension_Verification_Enabled == True \
		and fileExtension_Match_Found          == False:
			return []

		#▒▒▒  Populate Completions  ▒▒▒#

		completions = []

		for completionGroup in jsonCompletions:
			for completion in jsonCompletions[ completionGroup ]:
				completionText = completionGroup + completionSeparator + completion
				completions.append ( ( completionText, completionText ) )

		return ( completions )

 

And the two functions I use to find the last word & separator character:
( usage shown @ OP / GIF2 )
 

def get_LastWord ( view, R, wordOffset = 0 ):

	wordEnd_Position = R.column_B - wordOffset

	lastSpace_Postion = R.line_B_Text[ 0 : wordEnd_Position ].rfind ( " " )
	lastTab_Postion   = R.line_B_Text[ 0 : wordEnd_Position ].rfind ( "\t" )

	if lastSpace_Postion != -1:
		lastWord = R.line_B_Text[ lastSpace_Postion + 1 : wordEnd_Position ]
	elif lastTab_Postion != -1:
		lastWord = R.line_B_Text[ lastTab_Postion + 1 : wordEnd_Position ]
	else:
		lastWord = R.line_B_Text[ : wordEnd_Position ]

	return ( lastWord )

def get_LastCharacters ( view, R, count ):

	startPosition = R.column_B - count
	endPosition   = R.column_B

	if len ( R.line_B_Text ) < count:
		lastCharacters = R.line_B_Text[ : endPosition ]
	else:
		lastCharacters = R.line_B_Text[ startPosition : endPosition ]

	return ( lastCharacters )

 



 

 
Can you elaborate on why it would be useful to implement?

Personally, I’m good with the file extension filter.  I haven’t gotten into syntax & scopes all that much, and can’t think of a scenario where I would need a more precise filter.

However, I am considering adding it as a secondary filter, since it was included in the css_completions.py which I used as a starting point & it may be something others are interested in.

0 Likes

#4

The first example that comes to mind would be so it can offer autocompletion on unsaved files. But more importantly, some file types have many different “sub-languages” like a PHP file can contain HTML, JS and CSS as well as PHP, and you want to ensure you offer the correct/relevant completions in the right places.

1 Like

#5

it might be worth taking a look at how I did it in the XPath plugin. You can set which characters should automatically trigger auto completion, and then further down it makes use of the cursor positions to determine what the desired prefix (i.e. including word separator characters) is. Hopefully that will help you :wink:

1 Like

#6

 
Nice,

"auto_complete_triggers": [ {"selector": "text.html", "characters": "<"}, {"selector": "source.python", "characters": "."} ],

@ sublime-settings seemed to be what I needed.
 

 



 
It’s cool that you can set the triggers programmatically for the input panel ( I’ll have some fun with that another day :grin: ), but I’m guessing that sublime-settings/"auto_complete_triggers" is the only source for on_query_completion.
( couln’t find any other info… )

I was hoping to avoid necessary configuration on the user side, but at least it works.  Guess it’ll be on the devs to provide informative setup documentation.  Still not a big fan of that implementation since it means users will have to manage settings for multiple plugins in a single value.
 



 

 
Good point, I guess I’ll go with a 3-tier hierarchy:

  • Scope
  • Syntax
  • FileType

 
That way, lazy people like me can just throw a filetype or syntax in, and anyone who needs more control will have some options.

Also, I think FileType is equally as important as scope, because it allows you to define separate completions for different filetypes that may share the same syntax.

1 Like

#8

@kingkeith @FichteFoll

( & anyone else that’s interested :grin: )
 



 
I’ve revamped the code a good amount.   Here are the results:

 



 
Here’s what the .custom-completions files are looking like so far:

{
  "completionSeparator": ".",
  
  "completionMode": "Syntaxes", // "Global" | "Syntaxes" | "FileExtensions" | "Scopes"
  
  "completionGroup": [ "Packages/Python/Python.sublime-syntax" ],
  //■■■  EXAMPLES:  ■■■//
  // if completionMode == Global             : "completionGroup": [],
  // if completionMode == Syntaxes           : "completionGroup": [ "Packages/Java/Java.sublime-syntax", "Packages/Python/Python.sublime-syntax" ],
  // if completionMode == FileExtensions     : "completionGroup": [ "java", "py" ],
  // if completionMode == Scopes             : "completionGroup": [ "source.java", "source.python" ],
  // if completionMode == Scopes ( + RegEx ) : "completionGroup": [ "source\\.[^\\s]+" ],

  "completionSets":
    { //■■■  Completions : START  ■■■//

      "string": "THIS IS A STRING",

      "TestCompletion":
        {
          "PrimaryA": "PrimaryA",
          "PrimaryB": "PrimaryB",

          "subCompletions":
            {
              "PrimaryA":
                {
                  "SubA1": "SubA1",
                  "SubA2": "SubA2",
                },

              "PrimaryB":
                {
                  "SubB1": "SubB1",
                  "SubB2": "SubB2",
                },

              "PrimaryB.SubB2":
                {
                  "Nested1": "Nested1",
                  "Nested2": "Nested2",
                },
            }
        },

    } //■■■  Completions : END  ■■■//

}

 

Do you agree with my implementation of completionMode & completionGroup ?
 



 
The only completionMode that’s kind of tricky is scope

 
This seems to work fine, but I’d like your input just to be sure:

elif completionMode == completionMode_Scopes:

  for scope in completionGroup:
  
    rx_ScopeInScopes = \
        "(^" + scope + "($|\ ))" \   # matches entire scope or first scope
      + "|(.*\ " + scope + "($|\ ))" # matches last scope or any scope surrounded in spaces

    if re.match ( rx_ScopeInScopes, R.scope_B ): # R.scope_B is the scope @ caret region
      matchFound = True;
2 Likes

WebGL Completions and Tooltips
#9

looking good! there is a built in function view.match_selector which can be used for scope checking - that way, you can specify a scope like source.python - string, which will match if in a source.python scope but not in a string scope etc. Saves needing to use regex and keeps it consistent for developers expecting this functionality/behavior. :wink:

1 Like

#10

 
Cool, I’ll try it out tomorrow when I go back to work on it. :+1:

 

 
Something that is possible with my current implementation is a RegEx scope like:
source\\.[^\\s]+

Can you foresee any usage cases where that might come in handy, or should I completely ditch it & go with view.match_selector ?

0 Likes

#11

This can be emulated with the “selector” check using simply source.. And you can specify many different scopes like this: plain.text, source.python, so I would say ditch it. I do like the idea though - if this functionality didn’t exist, I would be singing your praises :wink:

1 Like

#12

I’m an avid YAML fan, so I’d recommend you to use that for human-created or -editable data. Additionally, I would abuse some “advanced” data structures to simplify it to the following.

completionSeparator: .

completionMode: Syntaxes # Global | Syntaxes | FileExtensions | Scopes

completionGroup: 
  - Packages/Python/Python.sublime-syntax

completionSets:
  - [string, THIS IS A STRING]
  - TestCompletion:
    - PrimaryA: 
      - SubA1
      - SubA2
    - PrimaryB: 
      - SubB1
      - SubB2:
        - Nested1
        - Nested2

I’m not sure how sub-completion discovery is supposed to work for completions where trigger and completion differ, but that is your problem to solve.

Furthermore, the “completion*” prefixes seem unnecessary since the file is, by extension, obviously already about completions. The term “group” is also misleading imo. I would rather use something like “syntax” or “context”.

I would also ditch the scopes+regex mode.

2 Likes

#13

 
My code currently tests only against the completions, since that’s really the only thing you can pull from the view.
 



 

 
Agreed.  I didn’t like that I was using completionGroup and completionSets, but I couldn’t think of the right word.

Context is perfect.  :+1:
 



 

 
I like how succinct it is, but I don’t like how non-visual it is.

For example:

      - SubB1
      - SubB2:
        - Nested1

In a larger document, a ( missing ) colon @ SubB2 could be a hassle to find if your completions aren’t working correctly; whereas in a JSON document, it’s pretty clear & easy to see what the hierarchy is.

I haven’t worked with YAML yet, so I’ll play around with it a bit & see how it goes.

 

Currently, all of my loops are explicit:

for file in completionFiles:

	for completions in completionSet:
		for completion in completions:

			for subCompletions in subCompletionSet:
				for subCompletion in subCompletions:

 
The YAML example you provided is more dynamic than my original JSON template; which removes redundancy from subCompletions, and I’m guessing would also allow an unlimited depth of nesting.

Can you offer some insight ( and/or pseudocode :grin: ) into how I could go about achieving a similar result without explicitly separating completions from subCompletions ?

0 Likes

#14

FYI

- SubB1
- SubB2:
  - Nested1
#
['SubB1', {'SubB2': ['Nested1']}]
- SubB1
- SubB2
  - Nested1
#
['SubB1', 'SubB2 - Nested1']

Indeed, that was also my intention.

No. I would probably provide some pseudo-code for my proposal, but not for this weird split into main and sub completions.
But maybe you can find something useful here (building the data structure) and here (crawling for completions).

0 Likes

#15

 

That’s what I meant.
( for the solution in your proposal )

My code for the completions & subCompletions approach is already 100% functional, as shown in the GIF demo above.

 

I haven’t worked with a recursive function of this sort, but I have a few ideas…

The one I’m leaning towards is splitting the lastWord @ caret region by completionSeparator, and digging into the ( single ) completions dictionary with each split string.
( or return the top level if no matches are found )

I’ll try it out when I resume working on this later today.

 

Thanks for your input & suggestions in the previous posts :slightly_smiling:

1 Like

#16

@FichteFoll



 
I spent the last 2 days trying to implement the YAML conversion.  While I was able to replicate the previously demonstrated level of functionality, I was unable to construct a working function to allow for arbitrary nesting.

I’m not too skilled with recursion yet, and in attempting to implement it I kind of hit a brick wall in regards to dealing with the complex nesting of arrays & dictionaries in YAML.

 

The closest I got was this script, which appends all nested entries from the YAML data to an array:

def run ( self, edit ):
...
	for entry in yamlData:
		self.get_AllEntries ( entry )
...
			
def get_AllEntries ( self, entry ):

	global entries

	if isinstance ( entry, dict ):

		dictKey = list ( entry.keys() ) [0]

		for key, value in entry.items():

			if  not isinstance ( value, dict ) \
			and not isinstance ( value, list ):
				entries.append ( ( key, value ) )

			elif isinstance ( value, dict ):
				self.get_AllEntries ( value )

			elif isinstance ( value, list ):
				entries.append ( dictKey )
				self.get_AllEntries ( value )

	elif isinstance ( entry, list ):
		for item in entry:
			self.get_AllEntries ( item )

	elif isinstance ( entry, str ):
		entries.append ( entry )

 

Although I was able to extract all of the YAML entries, it still didn’t help much as far as retrieving specific entries by key.
( especially performance-wise, considering my approach )
 



 
I’ve since reverted to my JSON model, and successfully managed to get recursive completions working with a pretty simple script which uses existing text to precisely query the JSON data:

# lastWord == all characters preceding caret until whitespace or BOL

# in the case of:
# lastWord == "TestCompletion.PrimaryB.SubB2"
# lastWord_Segment_Count == 3     ==     len ( lastWord.split ( "." ) )

	for index in range ( 0, lastWord_Segment_Count ):
		completions = completions[ lastWord_Segments[ index ] ]
	
	# is effectively:
	# completions[ TestCompletion ][ PrimaryB ][ SubB2 ]
	# which returns:
	# [ "Nested1", "Nested2" ]

 
Do you know of a simple way to apply a similar method to the YAML data?

My attempts so far have been pretty terrible performance-wise; I haven’t been able to get anywhere close to the efficiency shown in the JSON method above.

Considering the amount of potential completions, I definitely want to keep the filtering as minimal as possible.  Do you think it’s possible for a mixed array & dictionary YAML filter to perform as efficiently as an all dictionary JSON filter?

0 Likes

#17

Honestly, I don’t get your code. I simply do not understand what you want to accomplish. It seems like you are flattening the data structure into a single “list”, but that makes no sense because the nesting gets lost in the process and you can’t properly use it to provide context-specific completions.

(Besides, I find the way you are formatting your code and not following any consistent naming convention to be very confusing, but part of that is just me.)

I my proposal goes over your head, feel free to just not use it. I don’t currently have the time to write some sample code anyway.

0 Likes

#18

 
That’s the furthest functional code that I managed.  Beyond that, I tried integrating the

for index in range ( 0, lastWord_Segment_Count )

&

lastWord_Segments[ index ]

from my previous post in order to only append particular entries, but I wasn’t able to do it successfully beyond the first nested level.
 



 

 

Care to elaborate?

I’m not super versed in any particular language.  I pretty much just have the basics of programming down & am working on expanding my skillset through SublimeText, Android, & AHK projects ( for now ).  Def would apreciate any input :slightly_smiling:

0 Likes

#19

Check out PEP8, which includes the most widely known conventions for Python code. I personally do not follow all of them, but most. I especially dislike the whitespace characters around braces. Either way, variable names like lastWord_Segment_Count are a big no-no.

1 Like

#20

 
Is your issue with lastWord_Segment_Count based on understanding what it represents or formatting preference?
 



 
I just read through PEP8, and honestly I think it’s pretty heavily opinionated.

 

Examples:

Use 4 spaces per indentation level


Spaces are the preferred indentation method


Pet Peeves
Avoid extraneous whitespace in the following situations:

 
And also the use of K&R style indentation.

I personally prefer to read Lisp style, but use GNU as it is much easier to alter & maintain.

 

There are definitely some valid points within the guide, but many of them are purely based on personal preference & some are completely irrelevant due to the ability of editors like SublimeText to easily alter the mentioned preferences.

Combined with the fact that there’s a “Pet Peeves” section, it reads more like a rant than a style guide…
 



 

While I haven’t studied any particular language extensively, I have looked into general coding practices a fair amount.

The most helpful resource I’ve come across for this topic is:
Code Complete: A Practical Handbook of Software Construction

It’s takes a very logical approach, reviews many commonly used standards, and rather than specifying opinionated requirements - it offers unbiased explanations and comparisons.

 

Some examples that I personally have taken to are:

11.1 Considerations in Choosing Good Names

 

[Page 260]

The Most Important Naming Consideration

The most important consideration in naming a variable is that the name fully and
accurately describe the entity the variable represents. An effective technique for coming
up with a good name is to state in words what the variable represents. Often that
statement itself is the best variable name. It’s easy to read because it doesn’t contain
cryptic abbreviations, and it’s unambiguous. Because it’s a full description of the
entity, it won’t be confused with something else. And it’s easy to remember because
the name is similar to the concept.

For a variable that represents the number of people on the U.S. Olympic team, you
would create the name numberOfPeopleOnTheUsOlympicTeam. A variable that represents
the number of seats in a stadium would be numberOfSeatsInTheStadium. A variable
that represents the maximum number of points scored by a country’s team in any
modern Olympics would be maximumNumberOfPointsInModernOlympics. A variable
that contains the current interest rate is better named rate or interestRate than r or x.

 

[Page 274]

Format names to enhance readability

Two common techniques for increasing readability

are using capitalization and spacing characters to separate words. For example,
GYMNASTICSPOINTTOTAL is less readable than gymnasticsPointTotal or
gymnastics_point_total.

 

[Page 288]

Names should be as specific as possible. Names that are vague enough or general
enough to be used for more than one purpose are usually bad names.


Code is read far more times than it is written. Be sure that the names you choose
favor read-time convenience over write-time convenience.

 



 
On the topic of whitespace, I’d argue that it makes code much easier to read & scan through quickly.
( with less mental & visual processing effort )
 

For example:
Which of these is easier to parse for a specific rock?
 

[ 1 ]

 
[ 2 ]

 
If it’s just once, it’s not a big deal; but if I had to sort through 100s of groups on a daily basis - I’d most certainly go with the first style.

( I’m uber lazy when it comes to mental energy expenditure, gotta save my focus for actual coding problems  :stuck_out_tongue: )
 



 
I know there are a ton of debates on coding style, just thought I’d share some of the logic behind mine.
( along with some helpful resources, hopefully:slightly_smiling:

1 Like

#21

Thanks for the thorough reply.

re lastWord_Segment_Count: It mixes three different naming styles. camelCase, snake_case and TitleCase (for the underscore-separated words). Choosing good variable names does not matter if you can’t remember how you separated the sub-words from one another.
Python’s stdlib uses snake_case for most modules (cept the likes of unittest, which haven’t been migrated to the style yet), so using snake_case for your own code comes only naturally.

Python does not use K&R style because there are no braces for blocks in the syntax. You may use round braces to group statements.

That said, of course most of the conventions are going to be opinionated and some are not the best choices for each, but the most important aspect here is consistency. I read a lot of code and that code following the same style is a lot better for me than each programmer having his own take on “what is readable or not”. Even if the choices may be slightly sub-optimal, the fact that the style remains the same across code bases helps immensely.

I also do not agree with your stone example. The images show stones as the smalles entity (atomic), but code can be tokenized into groups, such as function calls, attribute or item access and treating those the same was as binary operands (surrounded by spaces) does not improve readability but makes it harder to grasp the extends of such tokens. To get back to your example: Think of it as stones of the same color or type being grouped while separated from the other groups and the task being “count the stones of each type”. Besides, you don’t wrap the accessor operator (.) with spaces either, do you?

But this is going off-topic now.

2 Likes

#22

 

My convention is pretty specific:
 

Class: title case
Person, HumanBeing

Variables: camel case + ( ( underscore + TitleCase ) * x )
item_Property_SubProperty, name, firstName, firstName_Vowel_Count

Functions: same as variables, but with a verb as the first word
verb_Description, get_FirstName
 

I only use camel case initially to differentiate between classes and [ variables | functions ], but find that title case offers the ability to minimize underscores while maintaining readability and also more clarity than strict camel case.

EG:
          lastWord_FirstSegment_Vowel_Count
vs      lastWordFirstSegmentVowelCount
vs      lastword_firstsegment_vowel_count
vs      last_word_first_segment_vowel_count

Although I will agree that this is my own take on “what is readable or not”, as you mentioned.
 

 
PEP8 shows mixed LISP and K&R style @ Indentation.

 

 
True.  I guess for in-line context, whitespace maybe detracts from code relationships.  I write more for macro-level scanning so I can find what I’m looking for quickly, and then utilize ST’s native bracket highlighting if I need to precisely interpret a particular line.  Also, I tend to over-declare variables as to avoid that issue entirely, so I can read complex nestings as if they were text rather than having to decipher functions in-line.

EG: I spread my rocks out so I can find them easily, and if I need to inspect one I’ll pick it up & stare at it.  :stuck_out_tongue:
 

 
Meh, I feel like it’s unavoidable since there are so many languages ( some of which have particular convention limitations ), potential styles, and personal preferences.

I know I’m not always going to dig everyone’s code preferences, but I can always use a tool to format their code or run a RegEx replace or something.  If I’m contributing to a repo I just try to adapt & use it as an exercise in versatility.

 

Have you used CoolFormat by any chance?  I find it super useful for quickly altering code that’s not in my preferred format.  Just wish it had a few more language options…  Know of any similar options that may be more universal?

Also, out of curiosity since you mentioned you review a lot of code - what kind of stuff do you work on outside of ST?

0 Likes