| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
|
| |
mainThread to avoid crashes if SP shows the data in the Content Browser in which the data will be imported.
[It occurred sometimes that the current table should be updated but no data were available due to threading.]
|
|
|
|
|
|
| |
values after re-invoking it
- store the accessory settings in SP's preferences
|
|
|
|
| |
- now the parser will be initialized by the import dialog's settings (was unintentionally deleted)
|
|
|
|
|
|
|
|
|
|
| |
functionality:
- Growls are now only shown by default if they are not fired from the frontmost window
- Long-running tasks (>3 secs) will still Growl
- Clicking on a Growl will now bring the associated window to the front
This addresses the original concerns of Issue #98.
|
|
|
|
|
|
| |
- Fix the Growl notification prefs message as well as making the dialog a sheet.
- Re-run genstrings to update localizable.strings and also remove use of multiple comments for a single string.
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
- Replace the CSV parsing function (arrayForCSV:) with a new SPCSVParser class
- Make speed improvements to SPCSVParser to achieve 1.9x faster parsing than the old arrayForCSV: function
- Rewrite CSV imports to be performed as a streaming import, keeping memory usage much much lower
- CSV field mapping preview is now shown very early on in the import process, as soon as the first hundred rwos are available for a preview
- Progress bars are more consistent and accurate
- CSV rows are grouped into batches of up to 50 (depending on line length) for import, falling back to one-query-per-row if errors occur. The current error reporting level is therefore maintained, but imports of non-erroring data are much much faster.
- Improve processing speed slightly
- Fix some odd edge cases in CSV parsing
This addresses issue #389.
|
|
|
|
| |
for confirmation, furthermore if a connection is available the user can choose 'Import' instead of loading it into the Query Editor for cases that an user invoked 'Open…' accidentally instead of 'Import…'
|
|
|
|
| |
DEFAULT NULL, but don't allow definition of tables as that - fix the export of temporary placeholder tables for those views. Thanks to Andreas Falk for report and details.
|
|
|
|
|
|
|
|
| |
- Rework CSV export to stream data, significantly reducing memory consumption and so increasing speed and stability when exporting large tables. By default safe/fast streaming is used, but a checkbox is available to select "low memory mode" full streaming, allowing export of any size table in theory. This addresses Issue #224.
- Rework XML export to stream data in the same way, also significantly reducing memory usage and providing the option of using low memory mode.
- Make SQL, CSV and XML export progress bars update more smoothly
- When exporting the current browse view or custom query result, show an indeterminate progress bar when copying large resultsets to avoid the app appearing to hang
|
|
|
|
|
|
| |
written to file [ this fixes issue 404 ]
• FieldEditorSheet: if string cell content is "NULL" select the entire string for convenience
|
|
|
|
|
|
|
| |
- Add defaults for fine-grained logging preferences
- Add a method to TableDocument to allow setting the query mode, and use the query mode to control logging
- Set import/export and custom query to set the appropriate query modes
|
|
|
|
|
|
|
| |
- SQL import now reads and processes files in full streaming mode, running queries as they are encountered
- Memory usage during import is significantly reduced, and should stay within a few megabytes; the significant memory use remaining is for query logging
- The progress bar more accurately represents progress and is shown at once (this addresses Issue #320)
|
|
|
|
| |
comments for a single string.
|
|
|
|
|
|
|
| |
download all results as fast as possible from the server, to avoid blocking, but do so in a background thread to allow results processing to start as soon as data is available. Many thanks to Hans-Jörg Bibiko for assistance with this.
- Add an option to the SQL export dialog to allow selection of the full-streaming method, with a warning that it may block table UPDATES/INSERTS.
|
|
|
|
|
|
| |
I had to cheat and name the file %@.csv, so the suggested filetype is csv, but you can change it to something else.
|
| |
|
| |
|
|
|
|
|
|
|
| |
- Added an MCPStreamingResult class to MCPKit, to allow streaming results from the server including fast array access of each row
- Tweak SQL export to use the streaming result class and to keep memory usage lower
End result is generally faster exports, more accurate progress bars, and much much lower (and consistent) memory usage.
|
|
|
|
|
|
|
|
|
| |
- Data loading now only occurs in one place in the code. This improves consistency and fixes a number of actions which used to trigger a full table reload followed instantly by a filter when the action was performed
- If "Reload data after..." prefs are unticked, no longer load the data (ie the preference now works)
- Make table count text more consistent and useful
- Fix a number of small position-saving type problems with filters and limits active. This fixes Issue #200.
- Clean up and standardise the code dealing with data storage - only one data storage array is now used.
|
| |
|
| |
|
|
|
|
|
|
|
| |
- When selecting CSVs SP will attempt to auto-detect the line endings to use
- Throw an error if the CSV to be imported appears to have more than 512 columns, usually due to wrong line ending/quote/etc selection
- Rewrite the CSV parser completely. New version correctly deals with CSV line terminators which are escaped or in enclosed content, correctly deals with quote strings which are the same as escape strings, and fixes a number of small edge cases. Performance on very long complex strings is slightly slower (~1.5 slower on long strings) but on large but simple tables is faster (~2.2x faster); memory usage is ~1.3x as high but all autoreleased. This addresses Issue #252.
|
| |
|
|
|
|
|
|
| |
- Enable threaded animation for import/export progress bars as it has minimal overhead and looks nicer
- Fix exceptions caused by tablesList drawing during imports
|
|
|
|
|
|
|
|
| |
- If there are twenty or more tables, show a table quicksearch/filter at the top of the list, and update the rest of the code to match. This addresses issue #178.
- Select tables and views alphabetically by user's current locale (instead of default MySQL "A B C a b c")
- When adding or duplicating tables, insert them at the correct point
- Fix a number of minor display bugs caused by incorrect interaction with the tables list caches
|
|
|
|
| |
autoreleased; this addresses Issue #341
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
- Includes all custom code from subclasses CMMCPConnection and CMMCPResult, meaning they have subsequently been removed from the project.
- All previous Sequel Pro specific code in the above subclasses has been removed in favour of the delegate (currently set to TableDocumet) informing the framework of such information.
- All references to CMMCPConnection and CMMCPResult have subsequently been changed to MCPConnection and MCPResult.
- Framework includes MySQL 5.1.36 client libraries and source headers.
- Framework is now built as a 4-way (32/64 bit, i386/PPC arch) binary.
- All import references to <MCPKit_bundled/MCPKit_bundled.h> have been changed to <MCPKit/MCPKit.h>.
- New script 'build-mysql-client.sh' can be used to build the MySQL client libraries from the MySQL source. See the script's header for a list of available options or run it with no arguments to display it's usage.
Note that there are still a few changes to be made to the framework with regard to removing Sequel Pro specific calls to the delegate. These however can be made later on as they have no effect on functionality and are merely design changes.
Also, note that any future development done on the framework should be made to be as 'generic' as possible, with no Sequel Pro specific references. This should allow the framework to be integrated into another project without the need for SP specific code.
|
|
|
|
|
|
|
|
|
|
| |
be closed when the window is closed, freeing the document's memory
- Update a number of dealloc methods to include more retained memory, and to avoid releasing non-retained memory
- Remove notification observers and delegates where appropriate to avoid issues after document closing
- Fix a couple of memory leaks
- Support window cascading for all windows past the first, using the first window as the autosave window
|
| |
|
|
|
|
|
| |
- Fix mutliple-table CSV and XML export when a view is selected - data for the view is now correctly exported
|
|
|
|
|
|
|
| |
and autorelease pools (fixes Issue #326)
- Reset the field mapping table list whenever it's about to be populated, fixing errors when performing imports into multi databases in the same session
|
|
|
|
| |
tables, resolving Issue #313
|
|
|
|
|
|
|
|
| |
- Since we are using mysql_real_query() there is no need to generate a \0 terminated cString.
- The disadvantage of using UTF8String is that it returns NULL if the encoding fails.
- To process each encoding equally we can now use [NSString dataUsingEncoding:enc allowLossyConversion:lossy]. In oder to avoid the overhead while calling it the following inline function is introduced:
NSStringDataUsingLossyEncoding(aStr, enc, lossy) := [aStr dataUsingEncoding:enc allowLossyConversion:lossy]
• import of CSV: code cleaned and optimized for speed a little
|
|
|
|
|
|
|
|
|
| |
- id o = NSArrayObjectAtIndex(anArray, index) :== id o = [anArray objectAtIndex:index]
- this speed up it ~3µs per call
- replaced that inline function for such calls within loops to speed up them
• used IMP function pointers for keepAlive calls within queryString:
• set -O3 (Fastest) compiler option
• allow in preference pane "Tables" to set the Limit up to 50000
|
|
|
|
| |
and views as a different header color
|
|
|
|
| |
fk relations
|
|
|
|
|
|
|
|
|
| |
statements in the Custom Query editor if necessary
- if statement begins with: use, create, alter, rename, drop
• reload table list, database pull-down menu according to imported statements
• sped up "Import MySQL Dump"
• fixed some tiny issues of the last commit
|
| |
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
- added
(NSArray *) splitSqlStringByCharacter:(unichar)character;
(NSArray *) splitSqlStringIntoRangesByCharacter:(unichar)character;
(long) firstOccurrenceInSqlOfCharacter:
to the SQLParser which recognize a “delimiter” command
• queryAtPosition now works with ranges to speed it up
- the current query ranges resp. the just activated query range are cached in order to avoid parsing if the user only navigates through the textView buffer, or if the user calls Run Prev/Current Query only
• the "import dump" function makes usage of that new “delimiter” support
- i.e. dumps with procs/funcs declaration could be imported
IMPORTANT: Please check the new SQLParser exhaustively in order to prove that new approach
|
| |
|
| |
|
|
|
|
|
|
| |
Also replaced some NSLog with DLog/ALog
|
|
|
|
|
|
| |
a bit more readable
- used for show/copy create view syntax as well as for a MySQL dump
|
|
|
|
| |
the sql file and was crashing due to the size of the file. Line is commented out in non-Debug builds.
|
|
|
|
|
|
|
|
| |
(issue #192).
- Also make these messages have a style of critical to indicate the potential loss of data as a result of performing the operation.
- Updated Localizable.strings to accommodate new dialog messages.
|
|
|
|
| |
remove '_dump' and spaces from filename.
|
|
|
|
| |
- committing Source
|
|
|
|
| |
queries are sent as UTF8, if possible; this correctly preserves encoding data when working with files export from Sequel Pro or via mysqldump, and attempts to fall back to the current connection encodings for other files. This should resolve Issue #116.
|